var/home/core/zuul-output/0000755000175000017500000000000015110343614014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110350075015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004026474615110350066017707 0ustar rootrootNov 22 14:27:12 crc systemd[1]: Starting Kubernetes Kubelet... Nov 22 14:27:13 crc restorecon[4699]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:13 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 22 14:27:14 crc restorecon[4699]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 22 14:27:14 crc kubenswrapper[4914]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 14:27:14 crc kubenswrapper[4914]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 22 14:27:14 crc kubenswrapper[4914]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 14:27:14 crc kubenswrapper[4914]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 14:27:14 crc kubenswrapper[4914]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 22 14:27:14 crc kubenswrapper[4914]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.596757 4914 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607200 4914 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607242 4914 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607252 4914 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607262 4914 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607270 4914 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607279 4914 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607288 4914 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607296 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607304 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607312 4914 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607321 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607329 4914 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607338 4914 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607345 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607353 4914 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607360 4914 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607368 4914 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607376 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607383 4914 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607391 4914 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607398 4914 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607406 4914 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607414 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607421 4914 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607431 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607438 4914 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607446 4914 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607454 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607461 4914 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607471 4914 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607480 4914 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607494 4914 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607507 4914 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607518 4914 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607528 4914 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607537 4914 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607547 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607557 4914 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607566 4914 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607576 4914 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607586 4914 feature_gate.go:330] unrecognized feature gate: Example Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607596 4914 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607609 4914 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607619 4914 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607628 4914 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607638 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607645 4914 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607655 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607662 4914 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607673 4914 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607683 4914 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607696 4914 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607706 4914 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607715 4914 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607724 4914 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607732 4914 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607739 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607747 4914 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607755 4914 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607763 4914 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607771 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607781 4914 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607789 4914 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607798 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607806 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607813 4914 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607821 4914 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607828 4914 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607836 4914 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607844 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.607851 4914 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608797 4914 flags.go:64] FLAG: --address="0.0.0.0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608822 4914 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608838 4914 flags.go:64] FLAG: --anonymous-auth="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608850 4914 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608862 4914 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608872 4914 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608913 4914 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608925 4914 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608934 4914 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608943 4914 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608953 4914 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608965 4914 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608975 4914 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608984 4914 flags.go:64] FLAG: --cgroup-root="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.608993 4914 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609002 4914 flags.go:64] FLAG: --client-ca-file="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609011 4914 flags.go:64] FLAG: --cloud-config="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609020 4914 flags.go:64] FLAG: --cloud-provider="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609029 4914 flags.go:64] FLAG: --cluster-dns="[]" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609481 4914 flags.go:64] FLAG: --cluster-domain="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609492 4914 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609501 4914 flags.go:64] FLAG: --config-dir="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609510 4914 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609520 4914 flags.go:64] FLAG: --container-log-max-files="5" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609533 4914 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609542 4914 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609551 4914 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609560 4914 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609570 4914 flags.go:64] FLAG: --contention-profiling="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609579 4914 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609590 4914 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609599 4914 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609609 4914 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609621 4914 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609630 4914 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609639 4914 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609648 4914 flags.go:64] FLAG: --enable-load-reader="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609657 4914 flags.go:64] FLAG: --enable-server="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609666 4914 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609679 4914 flags.go:64] FLAG: --event-burst="100" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609689 4914 flags.go:64] FLAG: --event-qps="50" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609698 4914 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609707 4914 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609716 4914 flags.go:64] FLAG: --eviction-hard="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609727 4914 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609737 4914 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609746 4914 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609757 4914 flags.go:64] FLAG: --eviction-soft="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609767 4914 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609776 4914 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609785 4914 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609794 4914 flags.go:64] FLAG: --experimental-mounter-path="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609803 4914 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609811 4914 flags.go:64] FLAG: --fail-swap-on="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609821 4914 flags.go:64] FLAG: --feature-gates="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609831 4914 flags.go:64] FLAG: --file-check-frequency="20s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609842 4914 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609851 4914 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609860 4914 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609869 4914 flags.go:64] FLAG: --healthz-port="10248" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609903 4914 flags.go:64] FLAG: --help="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609912 4914 flags.go:64] FLAG: --hostname-override="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609921 4914 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609930 4914 flags.go:64] FLAG: --http-check-frequency="20s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609939 4914 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609949 4914 flags.go:64] FLAG: --image-credential-provider-config="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609957 4914 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609967 4914 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609975 4914 flags.go:64] FLAG: --image-service-endpoint="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609984 4914 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.609994 4914 flags.go:64] FLAG: --kube-api-burst="100" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610003 4914 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610013 4914 flags.go:64] FLAG: --kube-api-qps="50" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610022 4914 flags.go:64] FLAG: --kube-reserved="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610031 4914 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610040 4914 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610050 4914 flags.go:64] FLAG: --kubelet-cgroups="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610059 4914 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610068 4914 flags.go:64] FLAG: --lock-file="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610077 4914 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610086 4914 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610095 4914 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610109 4914 flags.go:64] FLAG: --log-json-split-stream="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610119 4914 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610128 4914 flags.go:64] FLAG: --log-text-split-stream="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610137 4914 flags.go:64] FLAG: --logging-format="text" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610147 4914 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610156 4914 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610165 4914 flags.go:64] FLAG: --manifest-url="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610174 4914 flags.go:64] FLAG: --manifest-url-header="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610187 4914 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610196 4914 flags.go:64] FLAG: --max-open-files="1000000" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610207 4914 flags.go:64] FLAG: --max-pods="110" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610216 4914 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610225 4914 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610234 4914 flags.go:64] FLAG: --memory-manager-policy="None" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610244 4914 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610254 4914 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610263 4914 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610271 4914 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610291 4914 flags.go:64] FLAG: --node-status-max-images="50" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610335 4914 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610346 4914 flags.go:64] FLAG: --oom-score-adj="-999" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610355 4914 flags.go:64] FLAG: --pod-cidr="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610366 4914 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610381 4914 flags.go:64] FLAG: --pod-manifest-path="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610390 4914 flags.go:64] FLAG: --pod-max-pids="-1" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610399 4914 flags.go:64] FLAG: --pods-per-core="0" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610408 4914 flags.go:64] FLAG: --port="10250" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610417 4914 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610426 4914 flags.go:64] FLAG: --provider-id="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610435 4914 flags.go:64] FLAG: --qos-reserved="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610444 4914 flags.go:64] FLAG: --read-only-port="10255" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610453 4914 flags.go:64] FLAG: --register-node="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610462 4914 flags.go:64] FLAG: --register-schedulable="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610471 4914 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610487 4914 flags.go:64] FLAG: --registry-burst="10" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610496 4914 flags.go:64] FLAG: --registry-qps="5" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610504 4914 flags.go:64] FLAG: --reserved-cpus="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610515 4914 flags.go:64] FLAG: --reserved-memory="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610526 4914 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610535 4914 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610545 4914 flags.go:64] FLAG: --rotate-certificates="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610554 4914 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610562 4914 flags.go:64] FLAG: --runonce="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610571 4914 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610580 4914 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610590 4914 flags.go:64] FLAG: --seccomp-default="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610599 4914 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610608 4914 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610618 4914 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610627 4914 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610637 4914 flags.go:64] FLAG: --storage-driver-password="root" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610646 4914 flags.go:64] FLAG: --storage-driver-secure="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610655 4914 flags.go:64] FLAG: --storage-driver-table="stats" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610664 4914 flags.go:64] FLAG: --storage-driver-user="root" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610673 4914 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610683 4914 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610693 4914 flags.go:64] FLAG: --system-cgroups="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610701 4914 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610716 4914 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610724 4914 flags.go:64] FLAG: --tls-cert-file="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610735 4914 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610746 4914 flags.go:64] FLAG: --tls-min-version="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610755 4914 flags.go:64] FLAG: --tls-private-key-file="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610764 4914 flags.go:64] FLAG: --topology-manager-policy="none" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610773 4914 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610783 4914 flags.go:64] FLAG: --topology-manager-scope="container" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610793 4914 flags.go:64] FLAG: --v="2" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610804 4914 flags.go:64] FLAG: --version="false" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610816 4914 flags.go:64] FLAG: --vmodule="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610827 4914 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.610837 4914 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611075 4914 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611088 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611099 4914 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611108 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611117 4914 feature_gate.go:330] unrecognized feature gate: Example Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611125 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611133 4914 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611141 4914 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611159 4914 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611167 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611175 4914 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611183 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611191 4914 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611199 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611207 4914 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611215 4914 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611223 4914 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611230 4914 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611238 4914 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611245 4914 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611253 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611261 4914 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611269 4914 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611276 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611284 4914 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611291 4914 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611300 4914 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611307 4914 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611317 4914 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611327 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611336 4914 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611344 4914 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611354 4914 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611361 4914 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611369 4914 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611377 4914 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611385 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611393 4914 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611402 4914 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611410 4914 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611421 4914 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611429 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611437 4914 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611445 4914 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611453 4914 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611460 4914 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611468 4914 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611476 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611486 4914 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611496 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611505 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611513 4914 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611523 4914 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611530 4914 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611538 4914 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611547 4914 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611555 4914 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611564 4914 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611573 4914 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611581 4914 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611589 4914 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611598 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611608 4914 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611617 4914 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611625 4914 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611632 4914 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611640 4914 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611647 4914 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611655 4914 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611663 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.611670 4914 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.611743 4914 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.623969 4914 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.624022 4914 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624161 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624181 4914 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624190 4914 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624199 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624209 4914 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624221 4914 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624233 4914 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624242 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624250 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624258 4914 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624269 4914 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624279 4914 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624289 4914 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624298 4914 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624307 4914 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624316 4914 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624324 4914 feature_gate.go:330] unrecognized feature gate: Example Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624332 4914 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624340 4914 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624348 4914 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624356 4914 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624364 4914 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624371 4914 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624379 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624386 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624394 4914 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624401 4914 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624409 4914 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624416 4914 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624424 4914 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624432 4914 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624440 4914 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624447 4914 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624457 4914 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624468 4914 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624476 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624484 4914 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624492 4914 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624500 4914 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624507 4914 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624515 4914 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624523 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624530 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624538 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624545 4914 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624553 4914 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624561 4914 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624569 4914 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624577 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624584 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624592 4914 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624599 4914 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624607 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624615 4914 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624623 4914 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624631 4914 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624639 4914 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624646 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624900 4914 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624911 4914 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624918 4914 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624926 4914 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624934 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624941 4914 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624949 4914 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624957 4914 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624964 4914 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624975 4914 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624985 4914 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.624994 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625004 4914 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.625017 4914 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625254 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625268 4914 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625277 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625286 4914 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625294 4914 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625304 4914 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625315 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625325 4914 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625336 4914 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625346 4914 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625354 4914 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625362 4914 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625370 4914 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625379 4914 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625387 4914 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625395 4914 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625403 4914 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625410 4914 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625418 4914 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625425 4914 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625433 4914 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625442 4914 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625449 4914 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625456 4914 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625464 4914 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625472 4914 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625480 4914 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625487 4914 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625496 4914 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625503 4914 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625511 4914 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625520 4914 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625527 4914 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625535 4914 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625543 4914 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625551 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625559 4914 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625567 4914 feature_gate.go:330] unrecognized feature gate: Example Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625575 4914 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625583 4914 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625592 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625599 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625607 4914 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625615 4914 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625623 4914 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625630 4914 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625638 4914 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625646 4914 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625653 4914 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625660 4914 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625668 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625675 4914 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625683 4914 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625691 4914 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625701 4914 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625710 4914 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625719 4914 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625727 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625735 4914 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625743 4914 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625752 4914 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625760 4914 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625767 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625775 4914 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625783 4914 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625790 4914 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625798 4914 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625805 4914 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625813 4914 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625821 4914 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.625829 4914 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.625842 4914 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.626914 4914 server.go:940] "Client rotation is on, will bootstrap in background" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.632733 4914 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.632866 4914 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.634851 4914 server.go:997] "Starting client certificate rotation" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.634928 4914 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.635167 4914 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-25 11:07:17.712245648 +0000 UTC Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.635293 4914 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 68h40m3.076957545s for next certificate rotation Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.665495 4914 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.668791 4914 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.696130 4914 log.go:25] "Validated CRI v1 runtime API" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.736353 4914 log.go:25] "Validated CRI v1 image API" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.739003 4914 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.744068 4914 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-22-14-23-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.744115 4914 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.772311 4914 manager.go:217] Machine: {Timestamp:2025-11-22 14:27:14.770080983 +0000 UTC m=+0.578650704 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:eefbc55f-3f05-4b13-a0e3-46f39658fddc BootID:f45cf250-4ef3-489f-b51a-e18266c6e1d1 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:bc:82:8c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:bc:82:8c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b5:36:6e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5a:ac:c6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d4:0f:a6 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:48:a3:00 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b6:5e:6e:46:0b:9f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ee:47:74:53:9b:65 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.772691 4914 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.772920 4914 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.775075 4914 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.775408 4914 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.775477 4914 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.775809 4914 topology_manager.go:138] "Creating topology manager with none policy" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.775829 4914 container_manager_linux.go:303] "Creating device plugin manager" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.776460 4914 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.776528 4914 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.776954 4914 state_mem.go:36] "Initialized new in-memory state store" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.777092 4914 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.780332 4914 kubelet.go:418] "Attempting to sync node with API server" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.780365 4914 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.780413 4914 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.780436 4914 kubelet.go:324] "Adding apiserver pod source" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.780453 4914 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.784313 4914 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.785712 4914 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.787598 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.787860 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.787615 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.788180 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.788205 4914 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.790122 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.790279 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.790385 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.790544 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.790693 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791223 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791364 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791484 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791593 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791729 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791850 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.791992 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.792149 4914 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.792919 4914 server.go:1280] "Started kubelet" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.793292 4914 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.793699 4914 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.793864 4914 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.794787 4914 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 22 14:27:14 crc systemd[1]: Started Kubernetes Kubelet. Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.795569 4914 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.795612 4914 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.796000 4914 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 17:23:07.67757025 +0000 UTC Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.796077 4914 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 722h55m52.88149919s for next certificate rotation Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.796111 4914 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.796818 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.797025 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.796969 4914 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.796942 4914 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.797092 4914 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.798047 4914 factory.go:55] Registering systemd factory Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.798090 4914 factory.go:221] Registration of the systemd container factory successfully Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.798500 4914 server.go:460] "Adding debug handlers to kubelet server" Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.800771 4914 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.98:6443: connect: connection refused" interval="200ms" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.803854 4914 factory.go:153] Registering CRI-O factory Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.803934 4914 factory.go:221] Registration of the crio container factory successfully Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.804064 4914 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.804111 4914 factory.go:103] Registering Raw factory Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.804160 4914 manager.go:1196] Started watching for new ooms in manager Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.805687 4914 manager.go:319] Starting recovery of all containers Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.807384 4914 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.98:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a5a6e84fa079c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 14:27:14.792851356 +0000 UTC m=+0.601421057,LastTimestamp:2025-11-22 14:27:14.792851356 +0000 UTC m=+0.601421057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.821977 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822330 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822355 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822375 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822393 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822430 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822447 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822465 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822486 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822503 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822521 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822539 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822557 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822581 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822601 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822618 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822639 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822657 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822675 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822692 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822710 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822767 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822789 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822812 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822830 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822932 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822959 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822978 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.822996 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823014 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823032 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823049 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823075 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823092 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823110 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823127 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823145 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823162 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823181 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823198 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823215 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823232 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823250 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823277 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823296 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823313 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823333 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823350 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823369 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823391 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823409 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823434 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823458 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823479 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823499 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823519 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823537 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823557 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823573 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823591 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823608 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823625 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823642 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823660 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823679 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823696 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823714 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823731 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823747 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823766 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823783 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823800 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823817 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823834 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823850 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823870 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823914 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823931 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823948 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823964 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.823983 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824001 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824020 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824039 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824058 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824077 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824095 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824115 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824132 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824149 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824165 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824181 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824198 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824215 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824232 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824249 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824268 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824286 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824303 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824320 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824337 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824354 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824372 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824389 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824413 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824435 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824454 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824473 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824492 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824511 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824529 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824547 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824569 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824587 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824605 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824623 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824639 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824656 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824679 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824696 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824713 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824729 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824746 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824763 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824781 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824797 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824814 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824832 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824850 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824867 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824909 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824926 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824943 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824960 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824979 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.824997 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825013 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825030 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825047 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825066 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825082 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825099 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825118 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825135 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825160 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825178 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825195 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825211 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825232 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825250 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825273 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825291 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825309 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825327 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825345 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.825364 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828802 4914 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828845 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828868 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828921 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828946 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828966 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.828987 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829006 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829026 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829047 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829068 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829088 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829108 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829128 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829147 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829167 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829186 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829207 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829226 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829246 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829268 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829339 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829362 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829380 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829398 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829418 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829440 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829459 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829477 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829496 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829515 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829534 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829553 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829576 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829597 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829617 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829636 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829655 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829673 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829692 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829711 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829730 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829749 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829769 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829789 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829806 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829824 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829844 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829862 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829905 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829924 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829944 4914 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829964 4914 reconstruct.go:97] "Volume reconstruction finished" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.829976 4914 reconciler.go:26] "Reconciler: start to sync state" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.839049 4914 manager.go:324] Recovery completed Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.853386 4914 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.98:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a5a6e84fa079c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-22 14:27:14.792851356 +0000 UTC m=+0.601421057,LastTimestamp:2025-11-22 14:27:14.792851356 +0000 UTC m=+0.601421057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.858082 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.859860 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.859938 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.859952 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.860769 4914 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.860799 4914 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.860830 4914 state_mem.go:36] "Initialized new in-memory state store" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.863839 4914 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.866479 4914 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.866524 4914 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.866553 4914 kubelet.go:2335] "Starting kubelet main sync loop" Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.866602 4914 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 22 14:27:14 crc kubenswrapper[4914]: W1122 14:27:14.868214 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.868280 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.887275 4914 policy_none.go:49] "None policy: Start" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.888004 4914 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.888048 4914 state_mem.go:35] "Initializing new in-memory state store" Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.896323 4914 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.946590 4914 manager.go:334] "Starting Device Plugin manager" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.946938 4914 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.946966 4914 server.go:79] "Starting device plugin registration server" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.947538 4914 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.947567 4914 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.947767 4914 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.947931 4914 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.947954 4914 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 22 14:27:14 crc kubenswrapper[4914]: E1122 14:27:14.955505 4914 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.966663 4914 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.966813 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.968047 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.968088 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.968099 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.968280 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.968534 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.968632 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.969489 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.969548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.969561 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.969742 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.969922 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.969978 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971330 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971384 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971396 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971364 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971470 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971333 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971519 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971538 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971663 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971716 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.971744 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972365 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972413 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972430 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972614 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972674 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972684 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972810 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.972837 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973724 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973751 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973763 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973816 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973863 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973907 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973926 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.973952 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.975048 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.975093 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:14 crc kubenswrapper[4914]: I1122 14:27:14.975110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.001849 4914 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.98:6443: connect: connection refused" interval="400ms" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032563 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032624 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032660 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032734 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032791 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032840 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.032941 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033011 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033033 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033049 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033085 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033125 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033153 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033186 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.033215 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.048613 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.049968 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.050022 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.050041 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.050073 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.050681 4914 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.98:6443: connect: connection refused" node="crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.134976 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135273 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135377 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135517 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135553 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135584 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135614 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135646 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135673 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135711 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135676 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135736 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135773 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135770 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135810 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135839 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135849 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135854 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135804 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135921 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135927 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135862 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135896 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.135990 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.136019 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.136051 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.136051 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.136079 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.136124 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.136089 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.251311 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.252971 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.253030 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.253055 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.253093 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.253791 4914 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.98:6443: connect: connection refused" node="crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.313595 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.337854 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.350708 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.377833 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: W1122 14:27:15.385374 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e3917007578417fe5761108e273ce04326173a556a6f832b50d4b55295ef556c WatchSource:0}: Error finding container e3917007578417fe5761108e273ce04326173a556a6f832b50d4b55295ef556c: Status 404 returned error can't find the container with id e3917007578417fe5761108e273ce04326173a556a6f832b50d4b55295ef556c Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.387610 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 22 14:27:15 crc kubenswrapper[4914]: W1122 14:27:15.402163 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0a9d2de1c25334662550eda59a17f9a791e13bf58a6b6d6bbee763b9483cece5 WatchSource:0}: Error finding container 0a9d2de1c25334662550eda59a17f9a791e13bf58a6b6d6bbee763b9483cece5: Status 404 returned error can't find the container with id 0a9d2de1c25334662550eda59a17f9a791e13bf58a6b6d6bbee763b9483cece5 Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.402790 4914 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.98:6443: connect: connection refused" interval="800ms" Nov 22 14:27:15 crc kubenswrapper[4914]: W1122 14:27:15.409289 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-d7a03c827d6d27a486581a2e543733363b20abb11f808457b15d94382f398038 WatchSource:0}: Error finding container d7a03c827d6d27a486581a2e543733363b20abb11f808457b15d94382f398038: Status 404 returned error can't find the container with id d7a03c827d6d27a486581a2e543733363b20abb11f808457b15d94382f398038 Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.654256 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.656028 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.656089 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.656144 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.656192 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.656748 4914 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.98:6443: connect: connection refused" node="crc" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.793967 4914 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:15 crc kubenswrapper[4914]: W1122 14:27:15.841840 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.841946 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.870915 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e3917007578417fe5761108e273ce04326173a556a6f832b50d4b55295ef556c"} Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.872736 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ab797f7538082c535cccbc1ed805702ad5bcf5cf0d7161fb7ebb5849e69257f"} Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.873995 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"99f401a0c49bb422072f3ab5135f2176eb19bdd0bc46034dd3882662190d9d2b"} Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.875261 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d7a03c827d6d27a486581a2e543733363b20abb11f808457b15d94382f398038"} Nov 22 14:27:15 crc kubenswrapper[4914]: I1122 14:27:15.876243 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0a9d2de1c25334662550eda59a17f9a791e13bf58a6b6d6bbee763b9483cece5"} Nov 22 14:27:15 crc kubenswrapper[4914]: W1122 14:27:15.950311 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.950511 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:15 crc kubenswrapper[4914]: W1122 14:27:15.960968 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:15 crc kubenswrapper[4914]: E1122 14:27:15.961058 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:16 crc kubenswrapper[4914]: E1122 14:27:16.204151 4914 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.98:6443: connect: connection refused" interval="1.6s" Nov 22 14:27:16 crc kubenswrapper[4914]: W1122 14:27:16.297486 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:16 crc kubenswrapper[4914]: E1122 14:27:16.297585 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.456793 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.458562 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.458637 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.458676 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.458729 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:16 crc kubenswrapper[4914]: E1122 14:27:16.459482 4914 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.98:6443: connect: connection refused" node="crc" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.794193 4914 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.883322 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.883390 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.883410 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.883426 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.883464 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.884581 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.884615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.884630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.885421 4914 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6" exitCode=0 Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.885493 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.885513 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.886331 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.886366 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.886381 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.888047 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.888254 4914 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17" exitCode=0 Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.888346 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.888338 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.889176 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.889218 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.889235 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.889745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.889770 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.889784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.891801 4914 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4" exitCode=0 Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.891899 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.891926 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.893636 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.893707 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.893752 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.895320 4914 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1" exitCode=0 Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.895400 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1"} Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.895498 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.896742 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.896787 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:16 crc kubenswrapper[4914]: I1122 14:27:16.896804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.794690 4914 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:17 crc kubenswrapper[4914]: E1122 14:27:17.805439 4914 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.98:6443: connect: connection refused" interval="3.2s" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.903264 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.903298 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.904632 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.904804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.904966 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.908165 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.908390 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.908533 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.908230 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.910100 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.910133 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.910146 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.913370 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.913416 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.913439 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.913459 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.915397 4914 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502" exitCode=0 Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.915472 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502"} Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.915515 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.915581 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.916334 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.916369 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.916388 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.916939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.916994 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:17 crc kubenswrapper[4914]: I1122 14:27:17.917013 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.059796 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.060929 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.060960 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.060972 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.060995 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:18 crc kubenswrapper[4914]: E1122 14:27:18.061332 4914 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.98:6443: connect: connection refused" node="crc" Nov 22 14:27:18 crc kubenswrapper[4914]: W1122 14:27:18.121177 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:18 crc kubenswrapper[4914]: E1122 14:27:18.121268 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:18 crc kubenswrapper[4914]: W1122 14:27:18.129960 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:18 crc kubenswrapper[4914]: E1122 14:27:18.130015 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:18 crc kubenswrapper[4914]: W1122 14:27:18.417173 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:18 crc kubenswrapper[4914]: E1122 14:27:18.417252 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:18 crc kubenswrapper[4914]: W1122 14:27:18.712810 4914 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:18 crc kubenswrapper[4914]: E1122 14:27:18.713204 4914 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.98:6443: connect: connection refused" logger="UnhandledError" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.794017 4914 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.98:6443: connect: connection refused Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.924381 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1"} Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.924545 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.926777 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.926844 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.926866 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.928083 4914 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88" exitCode=0 Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.928186 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88"} Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.928549 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.928265 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.928947 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.929149 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.930118 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.930175 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.930188 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.930222 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.930239 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.930192 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.931303 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.931582 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:18 crc kubenswrapper[4914]: I1122 14:27:18.931731 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.778473 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.935415 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.935504 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.935936 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3"} Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.935983 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a"} Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.935995 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5"} Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.936802 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.936831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:19 crc kubenswrapper[4914]: I1122 14:27:19.936842 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.870988 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.944372 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.944429 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.945271 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e"} Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.945315 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913"} Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.945869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.945936 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.945952 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.946639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.946678 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:20 crc kubenswrapper[4914]: I1122 14:27:20.946695 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.262372 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.264602 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.264795 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.264993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.265223 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.685098 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.947081 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.947814 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.948008 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.948058 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.948085 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.948514 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.948542 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:21 crc kubenswrapper[4914]: I1122 14:27:21.948554 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.422983 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.423219 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.424556 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.424611 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.424630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.430145 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.949872 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.950005 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.950117 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.951621 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.951656 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.951665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.951673 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.951709 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:22 crc kubenswrapper[4914]: I1122 14:27:22.951729 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.190689 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.191016 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.192927 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.193008 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.193026 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.952126 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.953357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.953427 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:23 crc kubenswrapper[4914]: I1122 14:27:23.953448 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.081596 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.081839 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.083261 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.083313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.083331 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.920866 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.954041 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.955716 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.955773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:24 crc kubenswrapper[4914]: I1122 14:27:24.955797 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:24 crc kubenswrapper[4914]: E1122 14:27:24.956116 4914 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 22 14:27:25 crc kubenswrapper[4914]: I1122 14:27:25.368714 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:25 crc kubenswrapper[4914]: I1122 14:27:25.573982 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:25 crc kubenswrapper[4914]: I1122 14:27:25.957491 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:25 crc kubenswrapper[4914]: I1122 14:27:25.958813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:25 crc kubenswrapper[4914]: I1122 14:27:25.958866 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:25 crc kubenswrapper[4914]: I1122 14:27:25.958926 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.362506 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.362749 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.364180 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.364285 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.364310 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.960219 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.961694 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.961793 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:26 crc kubenswrapper[4914]: I1122 14:27:26.961814 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:28 crc kubenswrapper[4914]: I1122 14:27:28.574723 4914 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 14:27:28 crc kubenswrapper[4914]: I1122 14:27:28.574795 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 22 14:27:29 crc kubenswrapper[4914]: I1122 14:27:29.297142 4914 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 14:27:29 crc kubenswrapper[4914]: I1122 14:27:29.297250 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 14:27:29 crc kubenswrapper[4914]: I1122 14:27:29.308785 4914 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 22 14:27:29 crc kubenswrapper[4914]: I1122 14:27:29.308851 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.198824 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.199412 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.201199 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.201259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.201278 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.205970 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.980142 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.980275 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.981819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.981953 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:33 crc kubenswrapper[4914]: I1122 14:27:33.981976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.304455 4914 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.308861 4914 trace.go:236] Trace[580947479]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 14:27:22.248) (total time: 12060ms): Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[580947479]: ---"Objects listed" error: 12060ms (14:27:34.308) Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[580947479]: [12.060077428s] [12.060077428s] END Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.308936 4914 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.309343 4914 trace.go:236] Trace[1655402844]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 14:27:23.059) (total time: 11249ms): Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[1655402844]: ---"Objects listed" error: 11249ms (14:27:34.309) Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[1655402844]: [11.249903071s] [11.249903071s] END Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.309367 4914 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.310088 4914 trace.go:236] Trace[1165402574]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 14:27:23.202) (total time: 11107ms): Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[1165402574]: ---"Objects listed" error: 11107ms (14:27:34.309) Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[1165402574]: [11.107981648s] [11.107981648s] END Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.310130 4914 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.310422 4914 trace.go:236] Trace[1398685989]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Nov-2025 14:27:23.855) (total time: 10455ms): Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[1398685989]: ---"Objects listed" error: 10454ms (14:27:34.310) Nov 22 14:27:34 crc kubenswrapper[4914]: Trace[1398685989]: [10.455136227s] [10.455136227s] END Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.310448 4914 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.312817 4914 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.314059 4914 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.358117 4914 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45278->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.358131 4914 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45276->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.358189 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45278->192.168.126.11:17697: read: connection reset by peer" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.358293 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45276->192.168.126.11:17697: read: connection reset by peer" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.358710 4914 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.358745 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.792903 4914 apiserver.go:52] "Watching apiserver" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.797290 4914 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.797744 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.798289 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.798395 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.798480 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.798536 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.798587 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.799175 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.799225 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.799545 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.799767 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.802226 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.802285 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.802908 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.802953 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.803031 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.802961 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.803216 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.802922 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.803785 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.897972 4914 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.903333 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.920319 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.920610 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.920728 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.920824 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.920962 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921076 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921257 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921202 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921361 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921311 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921353 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921407 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921425 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921320 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921445 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921448 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921478 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921497 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921513 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921528 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921543 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921557 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921573 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921590 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921596 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921604 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921620 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921635 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921675 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921690 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921704 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921720 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921736 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921750 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921767 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921784 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921799 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921814 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921829 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921845 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921889 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921905 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921920 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921935 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921949 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921963 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921983 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921997 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922014 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922028 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922043 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922058 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922072 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922102 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922116 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922131 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922147 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922161 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922175 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922189 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922203 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922227 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922240 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922256 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922273 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922288 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922302 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922318 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922333 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922347 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.924229 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928901 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930815 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930859 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930901 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930918 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930939 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930959 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.930977 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921750 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921788 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.921903 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922035 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.922333 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928354 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928403 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928578 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928595 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928619 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928765 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.928780 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.929006 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.932943 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.933201 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.933381 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.933944 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.934295 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.934639 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935017 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935109 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935163 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935484 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935567 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935619 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935651 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935681 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935710 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935737 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935766 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935811 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935814 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935868 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935907 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935929 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935948 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935971 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.935994 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936012 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936031 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936052 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936071 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936076 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936091 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936129 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936159 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936189 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936224 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936259 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936286 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936314 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936345 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936371 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936373 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936398 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936440 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936476 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936516 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936547 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936576 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936609 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936637 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936690 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936715 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936723 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936743 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936771 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936795 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936821 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936848 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936889 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936917 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936945 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.936973 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937003 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937034 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937063 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937086 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937121 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937151 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937179 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937178 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937204 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937304 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937339 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937375 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937405 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937430 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937460 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937485 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937490 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937544 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937576 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937615 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937644 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937722 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937723 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937955 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.939785 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.940125 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.940322 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.940798 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.951698 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.951961 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.953340 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.953549 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.953601 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.953937 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.954140 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.955311 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.955569 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.955938 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.956198 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.956486 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.957030 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.957150 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.957615 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.957717 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958019 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958025 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.937756 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958331 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958367 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958393 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958422 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958446 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958471 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958496 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958534 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958560 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958587 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958611 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958635 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958657 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958682 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958704 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.958726 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.959464 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.959613 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.959653 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.959821 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.959849 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.960325 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.960352 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.960408 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.960740 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.960949 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961223 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961261 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961288 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961312 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961340 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961367 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961396 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961423 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961447 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961476 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961501 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961528 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961552 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961578 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961604 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961629 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961655 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961679 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961703 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961728 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961761 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961789 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.961814 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962006 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962039 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962065 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962089 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962112 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962139 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962163 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962186 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962212 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962235 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962261 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962287 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962311 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962336 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962363 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962450 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962483 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962554 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962599 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.962897 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.963481 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.963995 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.964454 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.964767 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.964986 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.965008 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.965401 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.965619 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968044 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968372 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968507 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968588 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968698 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968781 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968861 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968968 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969049 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969143 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969221 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969294 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969364 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969515 4914 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969578 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969641 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969705 4914 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969764 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969832 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969952 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970021 4914 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970085 4914 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970140 4914 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970196 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970254 4914 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970316 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970369 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970425 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970478 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970538 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970619 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970681 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970742 4914 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970802 4914 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970856 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971079 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971157 4914 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971238 4914 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971363 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971420 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971481 4914 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971544 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971598 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971656 4914 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971709 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971768 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971833 4914 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.971966 4914 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972040 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972103 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972156 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972208 4914 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972260 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972384 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972445 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972504 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972564 4914 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972626 4914 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972708 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972765 4914 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972839 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972932 4914 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973003 4914 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973064 4914 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973122 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973186 4914 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973249 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.975800 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968514 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.968854 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969111 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969474 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.969817 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970185 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970495 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.970965 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.972993 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973582 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973725 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.973899 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974121 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974146 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974522 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974539 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974634 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974918 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.974943 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.975000 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.975191 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.975380 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.975710 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.975949 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:27:35.47592607 +0000 UTC m=+21.284495731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.976103 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.976592 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.976603 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.977200 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.977426 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.977646 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.977742 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.977860 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:35.477838069 +0000 UTC m=+21.286407730 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.977997 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.978654 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.978668 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.978957 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.979114 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.979372 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.979750 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.979955 4914 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.980214 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.980423 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.980498 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.980650 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.980717 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:34 crc kubenswrapper[4914]: E1122 14:27:34.980749 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:35.480740832 +0000 UTC m=+21.289310493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981131 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981219 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.977698 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981390 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981451 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981516 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981580 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981395 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981480 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981568 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981643 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981727 4914 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981741 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981753 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981767 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981785 4914 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981784 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981795 4914 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981822 4914 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981836 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981849 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981858 4914 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981868 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981895 4914 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981907 4914 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981919 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981932 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981942 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981951 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981962 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981973 4914 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.981984 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982157 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982224 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982389 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982414 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982820 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982828 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982890 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.982965 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983027 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983120 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983246 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983291 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983499 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983542 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983631 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.983772 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984047 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984209 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984282 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984449 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984694 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984986 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.985054 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.985439 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.985345 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.985464 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.985908 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.986176 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.986258 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.986301 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.986559 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.986589 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.986743 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987029 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987279 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987209 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.984353 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987510 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987615 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987861 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.987900 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988022 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988072 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988035 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988161 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988417 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988550 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988775 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988860 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.988946 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.989194 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:34 crc kubenswrapper[4914]: I1122 14:27:34.989209 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.005743 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.006026 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.006049 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.006063 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.006131 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:35.506106854 +0000 UTC m=+21.314676515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.008616 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.008640 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.008652 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.008713 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:35.508676499 +0000 UTC m=+21.317246150 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:34.942691 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.016975 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.017074 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.019373 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.021784 4914 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1" exitCode=255 Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.021829 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1"} Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.024000 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.024229 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.024571 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.027499 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.027937 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.031326 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.032235 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.032622 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.034748 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.034975 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.035154 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.035407 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.036194 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.037144 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.037228 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.039075 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.040745 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.040893 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.044211 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.065048 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.066395 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.067539 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.077751 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.078362 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083471 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083549 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083614 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083628 4914 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083628 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083640 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083706 4914 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083712 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083719 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083731 4914 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083758 4914 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083772 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083782 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083792 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083802 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083812 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083822 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083831 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083841 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083850 4914 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083860 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083870 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083893 4914 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083903 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083912 4914 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083922 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083931 4914 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083942 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083951 4914 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083960 4914 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083973 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083983 4914 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.083993 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084004 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084013 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084023 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084032 4914 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084041 4914 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084051 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084060 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084069 4914 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084078 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084088 4914 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084100 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084110 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084118 4914 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084127 4914 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084136 4914 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084148 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084156 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084164 4914 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084173 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084185 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084196 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084206 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084217 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084227 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084235 4914 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084244 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084253 4914 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084262 4914 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084272 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084281 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084290 4914 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084302 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084311 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084321 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084335 4914 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084345 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084355 4914 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084365 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084375 4914 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084385 4914 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084393 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084402 4914 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084412 4914 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084420 4914 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084428 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084437 4914 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084447 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084455 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084465 4914 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084473 4914 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084482 4914 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084492 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084501 4914 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084510 4914 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084519 4914 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084529 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084538 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084546 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084555 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084564 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084573 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084583 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084592 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084601 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084612 4914 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084621 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084629 4914 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084639 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084648 4914 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084657 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084667 4914 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084676 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084685 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084695 4914 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084704 4914 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084713 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084721 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084742 4914 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084755 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084763 4914 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084771 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084780 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084789 4914 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084799 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.084808 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.090305 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.099743 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.108081 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.108150 4914 scope.go:117] "RemoveContainer" containerID="f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.109451 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.120437 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.122510 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.136080 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.142920 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.145380 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: W1122 14:27:35.150633 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cece1d14f37799da40cf43ff4f2ddaf2fc92427f0d30fdc9246baeb51a531ba2 WatchSource:0}: Error finding container cece1d14f37799da40cf43ff4f2ddaf2fc92427f0d30fdc9246baeb51a531ba2: Status 404 returned error can't find the container with id cece1d14f37799da40cf43ff4f2ddaf2fc92427f0d30fdc9246baeb51a531ba2 Nov 22 14:27:35 crc kubenswrapper[4914]: W1122 14:27:35.153462 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-982a81115be8834c98e2752803a41ce601e76ae540a0ded6058c37d49d50c677 WatchSource:0}: Error finding container 982a81115be8834c98e2752803a41ce601e76ae540a0ded6058c37d49d50c677: Status 404 returned error can't find the container with id 982a81115be8834c98e2752803a41ce601e76ae540a0ded6058c37d49d50c677 Nov 22 14:27:35 crc kubenswrapper[4914]: W1122 14:27:35.155811 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-21c43061a1b74ecc91df099d7852fd6019c3d8c230a977cc73485cde9192fd43 WatchSource:0}: Error finding container 21c43061a1b74ecc91df099d7852fd6019c3d8c230a977cc73485cde9192fd43: Status 404 returned error can't find the container with id 21c43061a1b74ecc91df099d7852fd6019c3d8c230a977cc73485cde9192fd43 Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.161905 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.169929 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.179825 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.192822 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.206180 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.487740 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.487807 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.487837 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.487924 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.487993 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:27:36.487957282 +0000 UTC m=+22.296526943 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.488040 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:36.488027994 +0000 UTC m=+22.296597765 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.488073 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.488204 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:36.488176187 +0000 UTC m=+22.296745838 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.579358 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.583529 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.588684 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.588749 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.588922 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.588960 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.588966 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.589042 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.589067 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.588975 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.589150 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:36.589122713 +0000 UTC m=+22.397692394 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.589185 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:36.589170934 +0000 UTC m=+22.397740825 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.595737 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.608045 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.609814 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.628167 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.643926 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.657920 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.667618 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.677906 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.688506 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.700086 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.709310 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.718287 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.729095 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.746656 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.758637 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.768645 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.867511 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.867699 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.904068 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-g7khd"] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.904425 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8hpjt"] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.904617 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.905061 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-f4twn"] Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.905096 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.905709 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:35 crc kubenswrapper[4914]: W1122 14:27:35.912860 4914 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 22 14:27:35 crc kubenswrapper[4914]: E1122 14:27:35.912921 4914 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.913200 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.913270 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.913938 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.914110 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.917746 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.918646 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.932448 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.932516 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.932616 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.932786 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.932950 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.933303 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992643 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-cni-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992706 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-k8s-cni-cncf-io\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992735 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-etc-kubernetes\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992758 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-kubelet\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992784 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-conf-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992806 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-proxy-tls\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992826 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-system-cni-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992841 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-socket-dir-parent\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992909 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98f7g\" (UniqueName: \"kubernetes.io/projected/1a5e8620-8854-4329-8935-a984f516d6b0-kube-api-access-98f7g\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992974 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-hostroot\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.992995 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-cni-binary-copy\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993010 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-cni-bin\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993034 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-os-release\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993049 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-multus-certs\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993073 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1a5e8620-8854-4329-8935-a984f516d6b0-hosts-file\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993102 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-mcd-auth-proxy-config\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993140 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5skzk\" (UniqueName: \"kubernetes.io/projected/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-kube-api-access-5skzk\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993194 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-cnibin\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993246 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-netns\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993301 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-daemon-config\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993355 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z46pd\" (UniqueName: \"kubernetes.io/projected/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-kube-api-access-z46pd\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993377 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-rootfs\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.993419 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-cni-multus\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:35 crc kubenswrapper[4914]: I1122 14:27:35.995101 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.026386 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.027832 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.028157 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.028779 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"21c43061a1b74ecc91df099d7852fd6019c3d8c230a977cc73485cde9192fd43"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.030281 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.030310 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.030320 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cece1d14f37799da40cf43ff4f2ddaf2fc92427f0d30fdc9246baeb51a531ba2"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.030530 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.031702 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.031726 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"982a81115be8834c98e2752803a41ce601e76ae540a0ded6058c37d49d50c677"} Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.048053 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.058592 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.070522 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.083246 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094385 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-cni-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094427 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-k8s-cni-cncf-io\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094449 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-etc-kubernetes\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094471 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-kubelet\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094492 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-conf-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094515 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-proxy-tls\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094534 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-k8s-cni-cncf-io\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094553 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-etc-kubernetes\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094551 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-system-cni-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094569 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-conf-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094613 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-socket-dir-parent\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094562 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-kubelet\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094636 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98f7g\" (UniqueName: \"kubernetes.io/projected/1a5e8620-8854-4329-8935-a984f516d6b0-kube-api-access-98f7g\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094668 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-hostroot\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094674 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-system-cni-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094708 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-os-release\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094712 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-hostroot\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094726 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-cni-binary-copy\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094727 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-socket-dir-parent\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094735 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-cni-dir\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094744 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-cni-bin\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094780 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-cni-bin\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094785 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-multus-certs\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094806 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-multus-certs\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094831 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1a5e8620-8854-4329-8935-a984f516d6b0-hosts-file\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094858 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-mcd-auth-proxy-config\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094899 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5skzk\" (UniqueName: \"kubernetes.io/projected/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-kube-api-access-5skzk\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094951 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-cnibin\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094957 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-os-release\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094998 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-netns\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095027 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-daemon-config\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095053 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-cnibin\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095035 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-run-netns\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.094993 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1a5e8620-8854-4329-8935-a984f516d6b0-hosts-file\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095098 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-cni-multus\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095124 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z46pd\" (UniqueName: \"kubernetes.io/projected/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-kube-api-access-z46pd\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095139 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-host-var-lib-cni-multus\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095146 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-rootfs\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.095196 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-rootfs\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.096647 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.114033 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.123620 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.135787 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.140801 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-cni-binary-copy\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.140857 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-proxy-tls\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.140993 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-mcd-auth-proxy-config\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.141247 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-multus-daemon-config\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.144526 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z46pd\" (UniqueName: \"kubernetes.io/projected/b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed-kube-api-access-z46pd\") pod \"multus-8hpjt\" (UID: \"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\") " pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.151249 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5skzk\" (UniqueName: \"kubernetes.io/projected/b8b64a09-b0a3-4776-93f7-01c4ea3e1388-kube-api-access-5skzk\") pod \"machine-config-daemon-g7khd\" (UID: \"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\") " pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.154844 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.171245 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.182781 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.193051 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.206397 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.218424 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.219615 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.229796 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8hpjt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.233079 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.246857 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.257447 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.271583 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: W1122 14:27:36.289092 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8b64a09_b0a3_4776_93f7_01c4ea3e1388.slice/crio-228aa5a02b9d0c4c25d97cc75171ec67731f317239aa91351ba377642d8a30d7 WatchSource:0}: Error finding container 228aa5a02b9d0c4c25d97cc75171ec67731f317239aa91351ba377642d8a30d7: Status 404 returned error can't find the container with id 228aa5a02b9d0c4c25d97cc75171ec67731f317239aa91351ba377642d8a30d7 Nov 22 14:27:36 crc kubenswrapper[4914]: W1122 14:27:36.289691 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1f94050_7d2c_4a11_9bb6_ce9c9cdaabed.slice/crio-b1e889337afd9fb08edbf2796b5bba8a4cb23c54c94967e50492fa5ebe3d15e6 WatchSource:0}: Error finding container b1e889337afd9fb08edbf2796b5bba8a4cb23c54c94967e50492fa5ebe3d15e6: Status 404 returned error can't find the container with id b1e889337afd9fb08edbf2796b5bba8a4cb23c54c94967e50492fa5ebe3d15e6 Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.322570 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jz8w8"] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.323789 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fftvp"] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.324911 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.327008 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.327249 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.331592 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.331702 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.331791 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.332085 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.332255 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.332411 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.332536 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.332674 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.341010 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.356075 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.372419 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.393205 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.395835 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.405399 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.407789 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.416699 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.433734 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.453094 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.475521 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.493287 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497678 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497757 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-netns\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497777 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-ovn\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497796 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-config\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497810 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-etc-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497823 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-log-socket\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497846 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-systemd-units\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497861 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497894 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-ovn-kubernetes\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497909 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-bin\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497923 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-script-lib\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497938 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-slash\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497953 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ea1017-213b-4be4-87ae-d68655d58a12-ovn-node-metrics-cert\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497969 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-var-lib-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.497997 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j94x\" (UniqueName: \"kubernetes.io/projected/41ea1017-213b-4be4-87ae-d68655d58a12-kube-api-access-6j94x\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498029 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498043 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cnibin\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498062 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498077 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-os-release\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498090 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-node-log\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498104 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-netd\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498118 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498138 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-kubelet\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498153 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-env-overrides\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498181 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498198 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-system-cni-dir\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.498215 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:27:38.498197935 +0000 UTC m=+24.306767596 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498248 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.498327 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.498348 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.498404 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:38.498389921 +0000 UTC m=+24.306959582 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.498418 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:38.498411712 +0000 UTC m=+24.306981363 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498434 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-systemd\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498451 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cni-binary-copy\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.498475 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f765\" (UniqueName: \"kubernetes.io/projected/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-kube-api-access-4f765\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.511507 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.527752 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.538887 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.553813 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.568819 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.583308 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.595279 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599432 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599477 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-systemd-units\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599503 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599528 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-ovn-kubernetes\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599548 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-bin\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599568 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-script-lib\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599595 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-slash\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599616 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ea1017-213b-4be4-87ae-d68655d58a12-ovn-node-metrics-cert\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.599632 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.599664 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.599678 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599681 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-var-lib-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599724 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-systemd-units\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.599728 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:38.599711536 +0000 UTC m=+24.408281197 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599762 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599795 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-ovn-kubernetes\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599824 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-bin\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.599637 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-var-lib-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600045 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j94x\" (UniqueName: \"kubernetes.io/projected/41ea1017-213b-4be4-87ae-d68655d58a12-kube-api-access-6j94x\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600071 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600087 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600103 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cnibin\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600134 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-os-release\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600151 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-node-log\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600165 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-netd\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600181 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600205 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-kubelet\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600243 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-env-overrides\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600259 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600283 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-system-cni-dir\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600300 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f765\" (UniqueName: \"kubernetes.io/projected/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-kube-api-access-4f765\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600320 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-systemd\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600336 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cni-binary-copy\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600354 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-netns\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600371 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-ovn\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600390 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-config\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600405 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-etc-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600419 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-log-socket\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600469 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-log-socket\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600623 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-script-lib\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.600644 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.600656 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.600664 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600673 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-slash\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.600689 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:38.60068055 +0000 UTC m=+24.409250211 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600714 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600737 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cnibin\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600783 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-os-release\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600803 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-node-log\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.600823 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-netd\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601216 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-systemd\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601277 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-kubelet\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601351 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601777 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-env-overrides\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601859 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601895 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-cni-binary-copy\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601909 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-system-cni-dir\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601937 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-netns\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.601962 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-ovn\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.602101 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-etc-openvswitch\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.602340 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-config\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.604084 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ea1017-213b-4be4-87ae-d68655d58a12-ovn-node-metrics-cert\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.606635 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.616152 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j94x\" (UniqueName: \"kubernetes.io/projected/41ea1017-213b-4be4-87ae-d68655d58a12-kube-api-access-6j94x\") pod \"ovnkube-node-fftvp\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.618124 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.620932 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f765\" (UniqueName: \"kubernetes.io/projected/d8eb63cc-74ad-43b4-9bf9-2cac458d53aa-kube-api-access-4f765\") pod \"multus-additional-cni-plugins-jz8w8\" (UID: \"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\") " pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.638826 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.651456 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.654543 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.656247 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:36 crc kubenswrapper[4914]: W1122 14:27:36.668832 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8eb63cc_74ad_43b4_9bf9_2cac458d53aa.slice/crio-75baff4cffab3e1fd7393e7836403066bb5a396f224deb7c98ce799f907261c4 WatchSource:0}: Error finding container 75baff4cffab3e1fd7393e7836403066bb5a396f224deb7c98ce799f907261c4: Status 404 returned error can't find the container with id 75baff4cffab3e1fd7393e7836403066bb5a396f224deb7c98ce799f907261c4 Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.669539 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: W1122 14:27:36.670108 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41ea1017_213b_4be4_87ae_d68655d58a12.slice/crio-e19ab2708505489d2bcef86f714e371f5c5d1f20bada0282c2c41b825e9ae44d WatchSource:0}: Error finding container e19ab2708505489d2bcef86f714e371f5c5d1f20bada0282c2c41b825e9ae44d: Status 404 returned error can't find the container with id e19ab2708505489d2bcef86f714e371f5c5d1f20bada0282c2c41b825e9ae44d Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.687694 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.707170 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.747075 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.783563 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.821968 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.867183 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.867206 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.867309 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:36 crc kubenswrapper[4914]: E1122 14:27:36.867487 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.871825 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.872518 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.873218 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.873852 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.874442 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.874964 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.875533 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.876102 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.876674 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.878633 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.879532 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.886343 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.887191 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.887737 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.888934 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.889580 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.890728 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.891225 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.891895 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.892946 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.893425 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.894402 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.894816 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.896046 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.896604 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.897358 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.898547 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.899256 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.900283 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.900841 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.901861 4914 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.902006 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.903781 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.904853 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.905374 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.906984 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.907633 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.908618 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.909334 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.910552 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.911159 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.912443 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.913238 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.914276 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.914733 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.916407 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.917063 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.918221 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.918722 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.919693 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.920157 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.921149 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.921726 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 22 14:27:36 crc kubenswrapper[4914]: I1122 14:27:36.922245 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.035955 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.036010 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.036023 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"228aa5a02b9d0c4c25d97cc75171ec67731f317239aa91351ba377642d8a30d7"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.037292 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" exitCode=0 Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.037367 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.037393 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"e19ab2708505489d2bcef86f714e371f5c5d1f20bada0282c2c41b825e9ae44d"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.038835 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerStarted","Data":"7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.038869 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerStarted","Data":"75baff4cffab3e1fd7393e7836403066bb5a396f224deb7c98ce799f907261c4"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.039780 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerStarted","Data":"6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.039808 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerStarted","Data":"b1e889337afd9fb08edbf2796b5bba8a4cb23c54c94967e50492fa5ebe3d15e6"} Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.063220 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.095856 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: E1122 14:27:37.115797 4914 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 22 14:27:37 crc kubenswrapper[4914]: E1122 14:27:37.115851 4914 projected.go:194] Error preparing data for projected volume kube-api-access-98f7g for pod openshift-dns/node-resolver-f4twn: failed to sync configmap cache: timed out waiting for the condition Nov 22 14:27:37 crc kubenswrapper[4914]: E1122 14:27:37.115932 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1a5e8620-8854-4329-8935-a984f516d6b0-kube-api-access-98f7g podName:1a5e8620-8854-4329-8935-a984f516d6b0 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:37.615912773 +0000 UTC m=+23.424482424 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-98f7g" (UniqueName: "kubernetes.io/projected/1a5e8620-8854-4329-8935-a984f516d6b0-kube-api-access-98f7g") pod "node-resolver-f4twn" (UID: "1a5e8620-8854-4329-8935-a984f516d6b0") : failed to sync configmap cache: timed out waiting for the condition Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.115907 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.137673 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.154185 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.161445 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.165097 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.177182 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.187577 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.207220 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.254249 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.288069 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.330789 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.366916 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.405084 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.443369 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.478459 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zpm9b"] Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.478803 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.486365 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.494712 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.516237 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.535396 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.554361 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.604071 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.609612 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz22n\" (UniqueName: \"kubernetes.io/projected/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-kube-api-access-lz22n\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.609668 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-serviceca\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.609705 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-host\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.647186 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.682335 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.711287 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-host\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.711336 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz22n\" (UniqueName: \"kubernetes.io/projected/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-kube-api-access-lz22n\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.711371 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-serviceca\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.711388 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98f7g\" (UniqueName: \"kubernetes.io/projected/1a5e8620-8854-4329-8935-a984f516d6b0-kube-api-access-98f7g\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.712122 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-host\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.713711 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-serviceca\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.717082 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98f7g\" (UniqueName: \"kubernetes.io/projected/1a5e8620-8854-4329-8935-a984f516d6b0-kube-api-access-98f7g\") pod \"node-resolver-f4twn\" (UID: \"1a5e8620-8854-4329-8935-a984f516d6b0\") " pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.723117 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.733672 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-f4twn" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.771039 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz22n\" (UniqueName: \"kubernetes.io/projected/7e6af9df-00d9-4baf-b5a5-dd22b5cc138e-kube-api-access-lz22n\") pod \"node-ca-zpm9b\" (UID: \"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\") " pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.784979 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.817401 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zpm9b" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.824484 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.867194 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:37 crc kubenswrapper[4914]: E1122 14:27:37.867325 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.872291 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.904338 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.946363 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:37 crc kubenswrapper[4914]: I1122 14:27:37.990575 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:37Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.028723 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.048318 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.048365 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.048375 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.048383 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.048392 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.049700 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.053799 4914 generic.go:334] "Generic (PLEG): container finished" podID="d8eb63cc-74ad-43b4-9bf9-2cac458d53aa" containerID="7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b" exitCode=0 Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.055924 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerDied","Data":"7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.055978 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zpm9b" event={"ID":"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e","Type":"ContainerStarted","Data":"9b7c0cf296c07b3ed4992e30d23c36ad1c737eddf3d946da91a6294bee50697a"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.057232 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-f4twn" event={"ID":"1a5e8620-8854-4329-8935-a984f516d6b0","Type":"ContainerStarted","Data":"e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.057256 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-f4twn" event={"ID":"1a5e8620-8854-4329-8935-a984f516d6b0","Type":"ContainerStarted","Data":"f0abd8e16ec32cca42c713baad930be5309ddcc2f23c494760ab19cb0242feca"} Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.060562 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.104793 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.157354 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.182420 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.224429 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.262981 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.300345 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.340426 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.381596 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.422364 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.463009 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.503366 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.518365 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.518516 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:27:42.518489848 +0000 UTC m=+28.327059509 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.518639 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.518679 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.518752 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.518793 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:42.518786207 +0000 UTC m=+28.327355868 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.518894 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.518994 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:42.518972651 +0000 UTC m=+28.327542332 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.549797 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.583749 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.619655 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.619790 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.619933 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.619969 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.619981 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.620040 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:42.620020209 +0000 UTC m=+28.428589870 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.620161 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.620192 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.620215 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.620305 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:42.620271575 +0000 UTC m=+28.428841266 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.623541 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.664324 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:38Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.866926 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:38 crc kubenswrapper[4914]: I1122 14:27:38.866965 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.867042 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:38 crc kubenswrapper[4914]: E1122 14:27:38.867199 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.063544 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.065236 4914 generic.go:334] "Generic (PLEG): container finished" podID="d8eb63cc-74ad-43b4-9bf9-2cac458d53aa" containerID="2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b" exitCode=0 Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.065324 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerDied","Data":"2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b"} Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.067090 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zpm9b" event={"ID":"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e","Type":"ContainerStarted","Data":"23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58"} Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.091408 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.104499 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.118464 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.130335 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.140972 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.160818 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.175899 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.187131 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.200476 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.214121 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.233910 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.245027 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.257675 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.267618 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.284527 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.304728 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.346224 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.381297 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.426646 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.463840 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.502825 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.545642 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.586191 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.625206 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.662837 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.703822 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.748196 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.791163 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.824846 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.865750 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:39Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:39 crc kubenswrapper[4914]: I1122 14:27:39.866850 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:39 crc kubenswrapper[4914]: E1122 14:27:39.867011 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.073405 4914 generic.go:334] "Generic (PLEG): container finished" podID="d8eb63cc-74ad-43b4-9bf9-2cac458d53aa" containerID="9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260" exitCode=0 Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.073502 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerDied","Data":"9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260"} Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.090223 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.104440 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.116868 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.134324 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.145089 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.153088 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.163252 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.182200 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.237717 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.277495 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.309269 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.342493 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.382736 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.421942 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.470121 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.713532 4914 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.717066 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.717121 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.717140 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.717261 4914 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.728218 4914 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.728590 4914 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.730026 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.730066 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.730078 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.730092 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.730103 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.748287 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.753333 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.753385 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.753403 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.753426 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.753444 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.771483 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.775810 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.775869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.775913 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.775934 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.775948 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.791575 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.797172 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.797250 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.797270 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.797298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.797318 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.819608 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.823488 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.823549 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.823566 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.823591 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.823608 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.843696 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:40Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.843994 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.845762 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.845812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.845829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.845850 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.845865 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.867570 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.867576 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.868314 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:40 crc kubenswrapper[4914]: E1122 14:27:40.868924 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.949154 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.949205 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.949222 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.949246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:40 crc kubenswrapper[4914]: I1122 14:27:40.949267 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:40Z","lastTransitionTime":"2025-11-22T14:27:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.052326 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.053033 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.053088 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.053115 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.053133 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.082851 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.087349 4914 generic.go:334] "Generic (PLEG): container finished" podID="d8eb63cc-74ad-43b4-9bf9-2cac458d53aa" containerID="474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c" exitCode=0 Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.087441 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerDied","Data":"474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.111619 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.135953 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.156707 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.156779 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.156803 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.156831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.156853 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.161699 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.190384 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.203604 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.216140 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.234689 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.259638 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.259688 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.259699 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.259717 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.259728 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.265129 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.287323 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.302294 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.318032 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.333033 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.347298 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.362156 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.363058 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.363095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.363107 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.363122 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.363503 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.377846 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:41Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.467399 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.467437 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.467448 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.467463 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.467474 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.571108 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.571154 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.571173 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.571198 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.571219 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.674419 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.674495 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.674516 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.674541 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.674560 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.778374 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.778412 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.778424 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.778444 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.778456 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.867314 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:41 crc kubenswrapper[4914]: E1122 14:27:41.867535 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.881711 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.881758 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.881781 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.881808 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.881829 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.985815 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.985861 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.985911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.985934 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:41 crc kubenswrapper[4914]: I1122 14:27:41.985951 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:41Z","lastTransitionTime":"2025-11-22T14:27:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.089620 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.089673 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.089685 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.089703 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.089714 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.095002 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerStarted","Data":"4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.119101 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.157739 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.182931 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.192449 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.192553 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.192571 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.192597 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.192615 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.207242 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.228530 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.244160 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.260324 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.279797 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.294594 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.296644 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.296712 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.296730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.296759 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.296777 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.312166 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.363447 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.395892 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.399625 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.399673 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.399683 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.399698 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.399708 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.413181 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.436609 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.453809 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.503495 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.503934 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.503958 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.503989 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.504008 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.562912 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.563114 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.563177 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.563292 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.563307 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:27:50.563270071 +0000 UTC m=+36.371839772 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.563385 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:50.563365253 +0000 UTC m=+36.371934914 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.563390 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.563529 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:50.563499876 +0000 UTC m=+36.372069577 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.606020 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.606087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.606104 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.606130 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.606149 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.663840 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.663955 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.664684 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.664793 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.664822 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.664923 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.664977 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.665003 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.665111 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:50.664936075 +0000 UTC m=+36.473505786 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.667319 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:27:50.667281455 +0000 UTC m=+36.475851136 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.709031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.709124 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.709151 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.709188 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.709214 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.812605 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.812638 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.812649 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.812665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.812677 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.866810 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.866973 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.867152 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:42 crc kubenswrapper[4914]: E1122 14:27:42.867431 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.915129 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.915173 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.915189 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.915209 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:42 crc kubenswrapper[4914]: I1122 14:27:42.915223 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:42Z","lastTransitionTime":"2025-11-22T14:27:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.018446 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.018512 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.018532 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.018559 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.018577 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.101649 4914 generic.go:334] "Generic (PLEG): container finished" podID="d8eb63cc-74ad-43b4-9bf9-2cac458d53aa" containerID="4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5" exitCode=0 Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.101711 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerDied","Data":"4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.108326 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.109031 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.109053 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.119441 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.121761 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.121796 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.121808 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.121824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.121841 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.138408 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.147155 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.147237 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.150073 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.162860 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.176132 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.189706 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.209710 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.225029 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.225070 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.225079 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.225093 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.225103 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.226757 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.242353 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.258258 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.271103 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.287525 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.309713 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.323498 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.327152 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.327200 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.327217 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.327239 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.327257 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.335801 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.345992 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.356668 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.367730 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.382397 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.402981 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.429482 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.429525 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.429535 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.429550 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.429568 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.431945 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.450271 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.465633 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.479167 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.495511 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.510150 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.525664 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.532055 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.532105 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.532115 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.532131 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.532142 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.536779 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.546168 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.553619 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:43Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.635636 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.635680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.635690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.635706 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.635717 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.738578 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.738642 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.738663 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.738691 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.738716 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.842447 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.842483 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.842491 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.842505 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.842542 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.867238 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:43 crc kubenswrapper[4914]: E1122 14:27:43.867349 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.947835 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.947901 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.947911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.947927 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:43 crc kubenswrapper[4914]: I1122 14:27:43.947935 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:43Z","lastTransitionTime":"2025-11-22T14:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.050670 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.050727 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.050738 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.050761 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.050774 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.115649 4914 generic.go:334] "Generic (PLEG): container finished" podID="d8eb63cc-74ad-43b4-9bf9-2cac458d53aa" containerID="2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992" exitCode=0 Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.115760 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerDied","Data":"2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.115861 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.146736 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.153350 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.153403 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.153421 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.153443 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.153463 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.164599 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.178946 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.204374 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.219792 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.234853 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.249241 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.256690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.257057 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.257312 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.257512 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.257743 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.264404 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.278025 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.296650 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.319925 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.342286 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.368997 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.369047 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.369064 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.369087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.369104 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.373853 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.388381 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.406522 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.471569 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.471603 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.471613 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.471628 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.471639 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.573864 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.573935 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.573951 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.573970 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.573981 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.676172 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.676208 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.676219 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.676235 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.676246 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.778465 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.778510 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.778521 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.778537 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.778549 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.867117 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.867206 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:44 crc kubenswrapper[4914]: E1122 14:27:44.867270 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:44 crc kubenswrapper[4914]: E1122 14:27:44.867355 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.880754 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.880783 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.880792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.880812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.880822 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.882253 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.893067 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.911849 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.922532 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.933803 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.946091 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.965507 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.979176 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.984504 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.984547 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.984558 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.984576 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.984587 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:44Z","lastTransitionTime":"2025-11-22T14:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:44 crc kubenswrapper[4914]: I1122 14:27:44.991865 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.005414 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.029207 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.045739 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.055512 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.069288 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.085168 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.089514 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.089553 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.089566 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.089580 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.089589 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.121762 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" event={"ID":"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa","Type":"ContainerStarted","Data":"4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.121835 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.134577 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.143805 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.152007 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.177782 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.189912 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.191353 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.191380 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.191390 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.191406 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.191417 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.206136 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.219023 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.229644 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.247587 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.268669 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.286670 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.293413 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.293448 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.293456 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.293470 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.293479 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.300640 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.313467 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.325817 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.334468 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.395652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.395719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.395734 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.395783 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.395798 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.497338 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.497400 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.497412 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.497429 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.497440 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.599988 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.600040 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.600056 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.600078 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.600093 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.702155 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.702185 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.702196 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.702210 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.702222 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.804079 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.804137 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.804146 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.804160 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.804169 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.867508 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:45 crc kubenswrapper[4914]: E1122 14:27:45.867710 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.906750 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.906794 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.906805 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.906824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:45 crc kubenswrapper[4914]: I1122 14:27:45.906842 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:45Z","lastTransitionTime":"2025-11-22T14:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.010021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.010063 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.010076 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.010096 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.010110 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.112316 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.112374 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.112392 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.112416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.112435 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.134122 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/0.log" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.139082 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21" exitCode=1 Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.139175 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.140364 4914 scope.go:117] "RemoveContainer" containerID="f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.159336 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.173970 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.189290 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.207291 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.219308 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.219633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.219662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.219745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.219820 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.224842 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.249737 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.281529 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.297680 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.312325 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.321630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.321666 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.321676 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.321693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.321705 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.330162 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.357698 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.375976 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.387645 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.400204 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.413913 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:46Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.423723 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.423754 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.423763 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.423777 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.423787 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.526665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.526707 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.526717 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.526756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.526768 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.629314 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.629362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.629376 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.629397 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.629414 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.731376 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.731415 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.731423 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.731438 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.731447 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.833756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.833791 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.833799 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.833813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.833822 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.867329 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.867396 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:46 crc kubenswrapper[4914]: E1122 14:27:46.867471 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:46 crc kubenswrapper[4914]: E1122 14:27:46.867583 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.936341 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.936379 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.936391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.936405 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:46 crc kubenswrapper[4914]: I1122 14:27:46.936416 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:46Z","lastTransitionTime":"2025-11-22T14:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.038332 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.038361 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.038369 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.038383 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.038391 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.140794 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.140824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.140831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.140844 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.140854 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.143293 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/0.log" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.145552 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.145655 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.158601 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.171471 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.190813 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.207869 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.218673 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.232993 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.244389 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.244593 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.244630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.244641 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.244657 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.244667 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.261808 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.273762 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.282562 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.294952 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.306189 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.317448 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.329457 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.340800 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:47Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.347792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.347821 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.347830 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.347891 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.347923 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.449628 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.449658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.449666 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.449678 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.449686 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.551981 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.552021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.552031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.552046 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.552057 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.654444 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.654484 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.654495 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.654510 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.654520 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.757617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.757679 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.757696 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.757720 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.757738 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.861997 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.862071 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.862094 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.862125 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.862146 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.867520 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:47 crc kubenswrapper[4914]: E1122 14:27:47.867745 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.964768 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.964834 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.964850 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.964908 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:47 crc kubenswrapper[4914]: I1122 14:27:47.965020 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:47Z","lastTransitionTime":"2025-11-22T14:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.067768 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.067825 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.067841 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.067865 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.067914 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.154529 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/1.log" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.155766 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/0.log" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.159827 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a" exitCode=1 Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.159930 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.160034 4914 scope.go:117] "RemoveContainer" containerID="f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.161376 4914 scope.go:117] "RemoveContainer" containerID="5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a" Nov 22 14:27:48 crc kubenswrapper[4914]: E1122 14:27:48.161733 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.170213 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.170290 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.170312 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.170339 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.170356 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.188451 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.205211 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.223339 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.241000 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.241490 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb"] Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.242289 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.244182 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.244197 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.268440 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.277839 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.277924 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.277945 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.277972 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.277991 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.295303 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.328301 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.337130 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms2wz\" (UniqueName: \"kubernetes.io/projected/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-kube-api-access-ms2wz\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.337279 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.337343 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.337383 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.347819 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.363385 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.377946 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.382390 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.382416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.382424 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.382437 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.382446 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.397984 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.414948 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.434791 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.437904 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.437940 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.437967 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.438011 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms2wz\" (UniqueName: \"kubernetes.io/projected/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-kube-api-access-ms2wz\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.439304 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.439538 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.447839 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.455541 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.468701 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms2wz\" (UniqueName: \"kubernetes.io/projected/be05ffa4-a006-46ca-996e-f5f4c4e0c42a-kube-api-access-ms2wz\") pod \"ovnkube-control-plane-749d76644c-mdlfb\" (UID: \"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.474570 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.485765 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.485822 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.485839 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.485865 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.485910 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.499287 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.522457 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.539431 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.554853 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.557531 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.574163 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: W1122 14:27:48.574589 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe05ffa4_a006_46ca_996e_f5f4c4e0c42a.slice/crio-aa749ca72330298c7b4ca7bc98620b96440eb6787f0df43ee0d84e2ba4cf2dc2 WatchSource:0}: Error finding container aa749ca72330298c7b4ca7bc98620b96440eb6787f0df43ee0d84e2ba4cf2dc2: Status 404 returned error can't find the container with id aa749ca72330298c7b4ca7bc98620b96440eb6787f0df43ee0d84e2ba4cf2dc2 Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.588790 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.588841 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.588859 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.589996 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.590134 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.593801 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.614925 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.632032 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.648394 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.659772 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.674354 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.692316 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.692343 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.692351 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.692363 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.692392 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.696702 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.711780 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.728809 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.743300 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.755847 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:48Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.796008 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.796033 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.796041 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.796053 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.796063 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.867058 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:48 crc kubenswrapper[4914]: E1122 14:27:48.867218 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.867378 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:48 crc kubenswrapper[4914]: E1122 14:27:48.867574 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.898851 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.898970 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.898993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.899027 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:48 crc kubenswrapper[4914]: I1122 14:27:48.899050 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:48Z","lastTransitionTime":"2025-11-22T14:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.001364 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.001404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.001416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.001435 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.001447 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.104225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.104286 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.104304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.104328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.104345 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.165997 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/1.log" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.183389 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" event={"ID":"be05ffa4-a006-46ca-996e-f5f4c4e0c42a","Type":"ContainerStarted","Data":"f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.183457 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" event={"ID":"be05ffa4-a006-46ca-996e-f5f4c4e0c42a","Type":"ContainerStarted","Data":"69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.183477 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" event={"ID":"be05ffa4-a006-46ca-996e-f5f4c4e0c42a","Type":"ContainerStarted","Data":"aa749ca72330298c7b4ca7bc98620b96440eb6787f0df43ee0d84e2ba4cf2dc2"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.202851 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.207156 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.207204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.207220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.207244 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.207261 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.224691 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.245809 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.263126 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.278977 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.299028 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.310736 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.310777 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.310794 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.310816 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.310834 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.318841 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.344656 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.368571 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.390356 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.404902 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.414439 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.414501 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.414517 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.414543 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.414560 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.419755 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.433223 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.450502 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.481847 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.502910 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.517763 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.517819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.517836 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.517864 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.517913 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.620765 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.620837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.620859 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.620914 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.620933 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.724091 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.724163 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.724181 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.724206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.724223 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.739650 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-47bbd"] Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.740574 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:49 crc kubenswrapper[4914]: E1122 14:27:49.740699 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.754642 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjz48\" (UniqueName: \"kubernetes.io/projected/d8b0b491-074a-465a-9419-6481d7877ecf-kube-api-access-jjz48\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.754753 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.774804 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.795644 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.822549 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.827630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.827805 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.827958 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.828085 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.828196 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.844727 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.855275 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjz48\" (UniqueName: \"kubernetes.io/projected/d8b0b491-074a-465a-9419-6481d7877ecf-kube-api-access-jjz48\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.855523 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:49 crc kubenswrapper[4914]: E1122 14:27:49.855775 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:49 crc kubenswrapper[4914]: E1122 14:27:49.855976 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:27:50.355940385 +0000 UTC m=+36.164510136 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.867216 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:49 crc kubenswrapper[4914]: E1122 14:27:49.867550 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.870138 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.888050 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjz48\" (UniqueName: \"kubernetes.io/projected/d8b0b491-074a-465a-9419-6481d7877ecf-kube-api-access-jjz48\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.891013 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.906331 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.922468 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.931054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.931120 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.931143 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.931173 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.931194 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:49Z","lastTransitionTime":"2025-11-22T14:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.938635 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.952851 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:49 crc kubenswrapper[4914]: I1122 14:27:49.980274 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.000413 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:49Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.016529 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.033466 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.033525 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.033534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.033547 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.033556 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.038719 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.057079 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.073367 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.092293 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.136808 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.136932 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.136962 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.137005 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.137029 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.240306 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.240393 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.240413 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.240438 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.240456 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.345031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.345096 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.345114 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.345138 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.345156 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.361800 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.362039 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.362141 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:27:51.362114618 +0000 UTC m=+37.170684309 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.447267 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.447297 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.447304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.447317 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.447326 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.550294 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.550336 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.550345 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.550362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.550372 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.590318 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.590530 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.590583 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:28:06.59053817 +0000 UTC m=+52.399107871 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.590653 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.590680 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.590782 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:06.590766867 +0000 UTC m=+52.399336568 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.590851 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.591019 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:06.590980442 +0000 UTC m=+52.399550143 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.653415 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.653479 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.653503 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.653530 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.653550 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.692107 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.692189 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692362 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692529 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692613 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692635 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692723 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:06.692697196 +0000 UTC m=+52.501266897 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692746 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692774 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.692844 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:06.692821489 +0000 UTC m=+52.501391190 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.756557 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.756663 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.756680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.756703 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.756719 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.860217 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.860278 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.860295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.860322 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.860337 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.866753 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.866816 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.866936 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.867076 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.878098 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.889115 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.890248 4914 scope.go:117] "RemoveContainer" containerID="5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.890646 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.892392 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.893387 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.893446 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.893466 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.893537 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.893556 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.906512 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.910705 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.914706 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.914748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.914764 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.914787 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.914808 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.920525 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.932680 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.936949 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.938210 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.938254 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.938270 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.938296 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.938312 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.953100 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.956017 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.959930 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.959993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.960012 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.960037 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.960055 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.974229 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: E1122 14:27:50.979287 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:50Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.985575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.985632 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.985651 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.985675 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:50 crc kubenswrapper[4914]: I1122 14:27:50.985696 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:50Z","lastTransitionTime":"2025-11-22T14:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: E1122 14:27:51.005773 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: E1122 14:27:51.006049 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.007725 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.007777 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.007795 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.007819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.007837 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.010621 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f324d52b1fdec64e2a2185b3b75c4d56ad64a514929d22c86a3bbb568fb59a21\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"message\\\":\\\"2 14:27:45.479023 6172 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1122 14:27:45.479051 6172 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:45.479061 6172 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1122 14:27:45.479126 6172 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1122 14:27:45.479139 6172 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1122 14:27:45.479162 6172 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1122 14:27:45.479176 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1122 14:27:45.479184 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1122 14:27:45.479192 6172 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:45.479200 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1122 14:27:45.479209 6172 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1122 14:27:45.479625 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:45.479640 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1122 14:27:45.479666 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:45.479690 6172 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1122 14:27:45.479707 6172 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.023542 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.051379 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.088487 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.107532 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.110298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.110336 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.110345 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.110359 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.110369 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.133376 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.147968 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.158060 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.168490 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.181236 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.192830 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.207045 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.212155 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.212192 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.212204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.212220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.212234 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.219985 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.232652 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.242231 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.252251 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.264668 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.278158 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.292967 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.314466 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.314922 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.314969 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.314980 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.314998 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.315013 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.343274 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.356342 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.366230 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.382580 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.395847 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.398910 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:51 crc kubenswrapper[4914]: E1122 14:27:51.399095 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:51 crc kubenswrapper[4914]: E1122 14:27:51.399174 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:27:53.399153391 +0000 UTC m=+39.207723062 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.409656 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.417654 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.417711 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.417731 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.417756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.417775 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.427584 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.441030 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:51Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.520677 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.520748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.520768 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.520797 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.520816 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.623310 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.623373 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.623393 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.623423 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.623445 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.726626 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.726683 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.726702 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.726723 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.726740 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.829457 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.829542 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.829561 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.829588 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.829607 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.867115 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.867228 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:51 crc kubenswrapper[4914]: E1122 14:27:51.867301 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:51 crc kubenswrapper[4914]: E1122 14:27:51.867441 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.932423 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.932516 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.932535 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.932560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:51 crc kubenswrapper[4914]: I1122 14:27:51.932578 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:51Z","lastTransitionTime":"2025-11-22T14:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.035485 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.035559 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.035584 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.035615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.035635 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.139188 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.139282 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.139309 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.139708 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.139728 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.242994 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.243060 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.243080 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.243112 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.243129 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.345634 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.345694 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.345713 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.345738 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.345756 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.448619 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.448685 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.448702 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.448728 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.448745 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.551264 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.551316 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.551333 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.551356 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.551374 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.654197 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.654262 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.654281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.654312 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.654333 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.757434 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.757476 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.757487 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.757504 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.757578 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.860229 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.860295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.860313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.860339 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.860356 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.867656 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.867723 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:52 crc kubenswrapper[4914]: E1122 14:27:52.867830 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:52 crc kubenswrapper[4914]: E1122 14:27:52.868048 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.964152 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.964364 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.964402 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.964434 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:52 crc kubenswrapper[4914]: I1122 14:27:52.964458 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:52Z","lastTransitionTime":"2025-11-22T14:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.068193 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.068259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.068276 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.068301 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.068318 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.172040 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.172166 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.172193 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.172220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.172238 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.274016 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.274066 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.274078 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.274096 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.274108 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.377053 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.377119 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.377138 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.377164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.377184 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.420331 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:53 crc kubenswrapper[4914]: E1122 14:27:53.420552 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:53 crc kubenswrapper[4914]: E1122 14:27:53.420651 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:27:57.420626324 +0000 UTC m=+43.229196025 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.479611 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.479678 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.479696 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.479721 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.479740 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.582667 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.582727 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.582746 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.582771 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.582788 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.685999 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.686074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.686096 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.686124 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.686148 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.788936 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.788976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.788988 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.789006 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.789018 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.867292 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.867353 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:53 crc kubenswrapper[4914]: E1122 14:27:53.867482 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:27:53 crc kubenswrapper[4914]: E1122 14:27:53.867619 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.891615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.891670 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.891686 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.891700 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.891709 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.993987 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.994049 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.994070 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.994096 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:53 crc kubenswrapper[4914]: I1122 14:27:53.994120 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:53Z","lastTransitionTime":"2025-11-22T14:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.097858 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.097972 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.097993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.098017 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.098035 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.213587 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.213647 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.213665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.213690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.213707 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.316665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.316730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.316746 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.316771 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.316787 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.419996 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.420067 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.420089 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.420120 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.420146 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.522966 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.523028 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.523046 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.523069 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.523088 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.626278 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.626344 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.626360 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.626386 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.626404 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.728710 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.728778 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.728799 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.728829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.728852 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.832006 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.832074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.832092 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.832118 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.832135 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.867814 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:54 crc kubenswrapper[4914]: E1122 14:27:54.867981 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.867815 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:54 crc kubenswrapper[4914]: E1122 14:27:54.868227 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.886619 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.905202 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.928606 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.934617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.934674 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.934692 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.934715 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.934733 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:54Z","lastTransitionTime":"2025-11-22T14:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.962249 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.979627 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:54 crc kubenswrapper[4914]: I1122 14:27:54.999141 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.018362 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.037768 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.038867 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.038957 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.038976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.039003 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.039023 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.072722 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.097314 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.115051 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.132974 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.141662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.141758 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.141813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.141841 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.141860 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.155807 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.177368 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.198139 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.216350 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.236625 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:27:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.244810 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.245058 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.245204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.245339 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.245469 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.349078 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.349433 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.349590 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.349746 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.349933 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.453645 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.453716 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.453742 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.453773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.453794 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.557312 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.557399 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.557417 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.557442 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.557460 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.660482 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.660526 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.660544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.660568 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.660586 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.764142 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.764193 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.764209 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.764232 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.764249 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867315 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:55 crc kubenswrapper[4914]: E1122 14:27:55.867561 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867616 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867703 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867722 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867736 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.867742 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:55 crc kubenswrapper[4914]: E1122 14:27:55.867999 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.970791 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.970825 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.970833 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.970846 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:55 crc kubenswrapper[4914]: I1122 14:27:55.970855 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:55Z","lastTransitionTime":"2025-11-22T14:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.074507 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.074572 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.074588 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.074612 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.074633 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.177380 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.177457 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.177474 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.177499 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.177516 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.280965 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.281044 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.281070 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.281099 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.281119 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.384987 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.385054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.385076 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.385113 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.385135 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.488436 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.488492 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.488508 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.488530 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.488548 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.590683 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.590740 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.590757 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.590806 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.590824 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.726455 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.726501 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.726517 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.726541 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.726559 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.828846 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.828939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.828957 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.828980 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.828997 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.867330 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:56 crc kubenswrapper[4914]: E1122 14:27:56.867504 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.867646 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:56 crc kubenswrapper[4914]: E1122 14:27:56.867830 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.932035 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.932097 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.932113 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.932137 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:56 crc kubenswrapper[4914]: I1122 14:27:56.932155 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:56Z","lastTransitionTime":"2025-11-22T14:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.035687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.035739 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.035755 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.035779 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.035796 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.139381 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.139447 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.139464 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.139487 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.139503 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.241735 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.241803 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.241820 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.241842 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.241861 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.344591 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.344634 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.344645 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.344662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.344679 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.447804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.447866 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.447915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.447940 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.447960 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: E1122 14:27:57.466656 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:57 crc kubenswrapper[4914]: E1122 14:27:57.466792 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:28:05.466761163 +0000 UTC m=+51.275330874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.466457 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.551143 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.551189 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.551206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.551232 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.551252 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.653152 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.653236 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.653259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.653288 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.653308 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.756752 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.756804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.756821 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.756840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.756856 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.859473 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.859543 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.859566 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.859597 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.859621 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.866738 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.866792 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:57 crc kubenswrapper[4914]: E1122 14:27:57.866945 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:27:57 crc kubenswrapper[4914]: E1122 14:27:57.867121 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.961647 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.961690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.961701 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.961716 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:57 crc kubenswrapper[4914]: I1122 14:27:57.961725 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:57Z","lastTransitionTime":"2025-11-22T14:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.064054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.064092 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.064101 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.064115 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.064124 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.166292 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.166361 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.166378 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.166402 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.166419 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.268633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.268674 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.268685 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.268701 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.268712 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.371383 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.371440 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.371456 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.371482 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.371502 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.474343 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.474394 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.474404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.474419 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.474431 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.577148 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.577219 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.577235 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.577261 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.577279 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.680144 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.680240 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.680259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.680287 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.680304 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.783639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.783687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.783704 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.783726 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.783742 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.867978 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:27:58 crc kubenswrapper[4914]: E1122 14:27:58.868163 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.868002 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:27:58 crc kubenswrapper[4914]: E1122 14:27:58.868278 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.886460 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.886509 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.886530 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.886556 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.886579 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.989406 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.989475 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.989565 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.989598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:58 crc kubenswrapper[4914]: I1122 14:27:58.989621 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:58Z","lastTransitionTime":"2025-11-22T14:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.092681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.092754 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.092776 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.092805 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.092827 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.195707 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.195767 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.195784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.195807 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.195826 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.299027 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.299098 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.299120 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.299207 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.299251 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.402281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.402347 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.402364 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.402385 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.402402 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.505765 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.505821 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.505839 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.505864 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.505907 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.609484 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.609530 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.609546 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.609567 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.609582 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.712808 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.712911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.712939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.712969 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.712988 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.815642 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.815712 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.815735 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.815762 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.815782 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.867362 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.867380 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:27:59 crc kubenswrapper[4914]: E1122 14:27:59.867555 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:27:59 crc kubenswrapper[4914]: E1122 14:27:59.867736 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.927226 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.927294 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.927311 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.927335 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:27:59 crc kubenswrapper[4914]: I1122 14:27:59.927351 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:27:59Z","lastTransitionTime":"2025-11-22T14:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.030798 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.030839 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.030847 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.030861 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.030869 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.133538 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.133573 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.133583 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.133598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.133611 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.236681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.236719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.236730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.236743 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.236754 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.339794 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.339861 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.339894 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.339937 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.339950 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.442481 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.442548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.442574 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.442604 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.442626 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.544987 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.545056 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.545078 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.545106 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.545137 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.648628 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.648693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.648711 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.648736 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.648755 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.752448 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.752506 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.752523 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.752544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.752558 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.854849 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.854967 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.854997 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.855026 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.855050 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.867622 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.867646 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:00 crc kubenswrapper[4914]: E1122 14:28:00.867846 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:00 crc kubenswrapper[4914]: E1122 14:28:00.868072 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.958546 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.958626 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.958645 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.958676 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:00 crc kubenswrapper[4914]: I1122 14:28:00.958696 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:00Z","lastTransitionTime":"2025-11-22T14:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.031689 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.031759 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.031777 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.031803 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.031821 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.053261 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:01Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.058440 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.058510 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.058536 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.058575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.058603 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.080438 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:01Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.085719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.085820 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.085843 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.085872 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.085927 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.108102 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:01Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.113929 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.113989 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.114009 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.114038 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.114061 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.142468 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:01Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.148627 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.148688 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.148709 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.148737 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.148754 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.168643 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:01Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.168859 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.171000 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.171060 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.171076 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.171102 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.171122 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.275189 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.275263 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.275285 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.275315 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.275340 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.378980 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.379098 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.379119 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.379146 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.379165 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.482266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.482341 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.482361 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.482387 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.482404 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.586553 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.586679 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.586701 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.586729 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.586748 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.690802 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.690872 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.690913 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.690945 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.690964 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.794827 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.794925 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.794944 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.794973 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.794992 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.867746 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.867940 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.867962 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:01 crc kubenswrapper[4914]: E1122 14:28:01.868510 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.898528 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.898593 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.898617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.898643 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:01 crc kubenswrapper[4914]: I1122 14:28:01.898664 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:01Z","lastTransitionTime":"2025-11-22T14:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.001197 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.001246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.001264 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.001288 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.001317 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.104127 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.104186 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.104203 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.104224 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.104241 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.207808 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.207909 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.207928 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.207951 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.207968 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.312243 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.312310 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.312328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.312354 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.312374 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.415304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.415375 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.415395 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.415417 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.415435 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.521018 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.521087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.521097 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.521113 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.521123 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.623512 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.623554 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.623567 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.623586 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.623596 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.726240 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.726310 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.726327 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.726351 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.726394 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.829629 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.829682 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.829696 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.829715 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.829730 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.866846 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.866904 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:02 crc kubenswrapper[4914]: E1122 14:28:02.867100 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:02 crc kubenswrapper[4914]: E1122 14:28:02.867190 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.932108 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.932141 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.932149 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.932166 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:02 crc kubenswrapper[4914]: I1122 14:28:02.932176 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:02Z","lastTransitionTime":"2025-11-22T14:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.035367 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.035494 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.035523 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.035559 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.035582 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.139031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.139110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.139128 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.139157 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.139178 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.247443 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.247526 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.247551 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.247592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.247617 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.351457 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.351519 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.351536 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.351589 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.351609 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.454760 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.454838 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.454860 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.455580 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.455666 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.559302 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.559366 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.559382 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.559404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.559419 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.665222 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.665305 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.665329 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.665361 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.665384 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.770074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.770145 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.770164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.770194 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.770213 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.867776 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.867816 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:03 crc kubenswrapper[4914]: E1122 14:28:03.868133 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:03 crc kubenswrapper[4914]: E1122 14:28:03.868337 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.874562 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.874617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.874637 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.874662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.874681 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.978413 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.978478 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.978496 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.978522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:03 crc kubenswrapper[4914]: I1122 14:28:03.978540 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:03Z","lastTransitionTime":"2025-11-22T14:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.082947 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.083031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.083054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.083087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.083110 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.090042 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.106566 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.114144 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.134436 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.151240 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.174460 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.187395 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.187502 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.187521 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.187548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.187570 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.195830 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.218408 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.262807 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.286648 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.290344 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.290396 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.290413 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.290437 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.290452 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.317867 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.333830 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.358527 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.377831 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.396245 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.396319 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.396343 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.396377 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.396402 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.397822 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.415347 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.436774 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.451765 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.468827 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.499618 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.499677 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.499695 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.499719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.499735 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.603693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.603792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.603813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.603840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.603858 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.706527 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.706575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.706592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.706616 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.706631 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.809505 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.809551 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.809562 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.809584 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.809597 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.867841 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.868401 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:04 crc kubenswrapper[4914]: E1122 14:28:04.868492 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:04 crc kubenswrapper[4914]: E1122 14:28:04.868686 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.893336 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.911234 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.913507 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.913549 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.913566 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.913588 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.913606 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:04Z","lastTransitionTime":"2025-11-22T14:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.927807 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.942255 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.956667 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.970333 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:04 crc kubenswrapper[4914]: I1122 14:28:04.984990 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:04Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.011592 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.017162 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.017193 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.017204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.017220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.017230 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.028678 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.044166 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.056663 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.075163 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.091708 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.115314 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.120717 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.120804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.120823 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.120860 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.120907 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.145349 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.164563 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.179338 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.200734 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:05Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.224043 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.224111 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.224127 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.224147 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.224160 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.326313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.326377 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.326396 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.326423 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.326442 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.429062 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.429122 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.429141 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.429164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.429181 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.531765 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.531816 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.531829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.531847 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.531860 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.562627 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:05 crc kubenswrapper[4914]: E1122 14:28:05.562803 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:28:05 crc kubenswrapper[4914]: E1122 14:28:05.562934 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:28:21.562907746 +0000 UTC m=+67.371477437 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.636534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.636606 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.636624 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.636648 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.636665 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.740237 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.740291 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.740308 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.740329 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.740347 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.843288 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.843351 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.843368 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.843392 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.843410 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.866907 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.866981 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:05 crc kubenswrapper[4914]: E1122 14:28:05.867099 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:05 crc kubenswrapper[4914]: E1122 14:28:05.867663 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.868254 4914 scope.go:117] "RemoveContainer" containerID="5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.946962 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.947005 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.947022 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.947047 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:05 crc kubenswrapper[4914]: I1122 14:28:05.947064 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:05Z","lastTransitionTime":"2025-11-22T14:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.050015 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.050259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.050270 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.050288 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.050300 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.152340 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.152380 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.152392 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.152409 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.152422 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.255485 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.255534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.255543 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.255559 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.255567 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.266719 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/1.log" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.269768 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.270332 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.287826 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.310379 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.345490 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.363739 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.363794 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.363814 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.363840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.363860 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.369692 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.386056 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.402912 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.431159 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.455084 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.466811 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.466851 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.466862 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.466903 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.466916 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.474974 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.489678 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.511978 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.523690 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.533911 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.548235 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.565091 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.569733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.569773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.569783 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.569799 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.569810 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.576293 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.591200 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.604894 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:06Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672397 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672455 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672469 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672489 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672503 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672822 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.672993 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.673035 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.673120 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.673177 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:38.673157462 +0000 UTC m=+84.481727133 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.673767 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:28:38.673752276 +0000 UTC m=+84.482321947 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.673860 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.673934 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:38.67391963 +0000 UTC m=+84.482489301 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.773825 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.773938 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774060 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774097 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774116 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774113 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774144 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774163 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774194 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:38.774173419 +0000 UTC m=+84.582743090 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.774218 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:28:38.77420985 +0000 UTC m=+84.582779521 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.775330 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.775400 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.775411 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.775425 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.775436 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.867448 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.867480 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.867658 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:06 crc kubenswrapper[4914]: E1122 14:28:06.867780 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.878047 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.878095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.878107 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.878126 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.878141 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.980156 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.980194 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.980203 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.980218 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:06 crc kubenswrapper[4914]: I1122 14:28:06.980227 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:06Z","lastTransitionTime":"2025-11-22T14:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.082751 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.082792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.082800 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.082815 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.082829 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.185222 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.185269 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.185281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.185298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.185309 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.275145 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/2.log" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.275854 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/1.log" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.279322 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb" exitCode=1 Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.279367 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.279402 4914 scope.go:117] "RemoveContainer" containerID="5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.280123 4914 scope.go:117] "RemoveContainer" containerID="6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb" Nov 22 14:28:07 crc kubenswrapper[4914]: E1122 14:28:07.280298 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.287048 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.287071 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.287079 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.287092 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.287100 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.307736 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.328148 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.343416 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.358465 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.378033 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.389624 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.389656 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.389668 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.389701 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.389712 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.393967 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.406292 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.420610 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.433617 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.447540 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.461010 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.483475 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.493920 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.494018 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.494039 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.494066 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.494084 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.506350 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a0c480ef3ec121996c393627f7b43a6fe03691abb6b1b76593232b7792d2b9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:27:47Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1122 14:27:47.050147 6317 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050289 6317 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.050706 6317 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1122 14:27:47.056866 6317 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1122 14:27:47.056922 6317 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1122 14:27:47.056937 6317 factory.go:656] Stopping watch factory\\\\nI1122 14:27:47.056961 6317 handler.go:208] Removed *v1.Node event handler 2\\\\nI1122 14:27:47.056962 6317 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1122 14:27:47.067088 6317 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1122 14:27:47.067115 6317 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1122 14:27:47.067182 6317 ovnkube.go:599] Stopped ovnkube\\\\nI1122 14:27:47.067200 6317 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:27:47.067272 6317 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.520857 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.550849 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.571177 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.587618 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.597011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.597066 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.597082 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.597107 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.597123 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.605037 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:07Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.699939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.699989 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.700009 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.700033 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.700052 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.803730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.804419 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.804449 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.804522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.804549 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.867316 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.867367 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:07 crc kubenswrapper[4914]: E1122 14:28:07.867867 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:07 crc kubenswrapper[4914]: E1122 14:28:07.867864 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.908011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.908299 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.908483 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.908648 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:07 crc kubenswrapper[4914]: I1122 14:28:07.908772 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:07Z","lastTransitionTime":"2025-11-22T14:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.011704 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.011753 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.011770 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.011792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.011811 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.115674 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.115730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.115746 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.115769 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.115787 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.219162 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.219231 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.219253 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.219281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.219302 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.285212 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/2.log" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.290269 4914 scope.go:117] "RemoveContainer" containerID="6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb" Nov 22 14:28:08 crc kubenswrapper[4914]: E1122 14:28:08.290522 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.311207 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.324220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.324273 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.324293 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.324316 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.324333 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.329686 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.346573 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.366693 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.386015 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.406087 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.427114 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.428367 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.428414 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.428431 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.428455 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.428474 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.450836 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.483824 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.500491 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.531943 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.532011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.532034 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.532063 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.532086 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.534068 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.556473 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.574743 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.594934 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.619396 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.634525 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.634568 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.634580 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.634596 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.634608 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.641241 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.659640 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.677213 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:08Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.738439 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.738512 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.738534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.738563 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.738586 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.841218 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.841264 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.841280 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.841301 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.841317 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.866947 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.866968 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:08 crc kubenswrapper[4914]: E1122 14:28:08.867135 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:08 crc kubenswrapper[4914]: E1122 14:28:08.867248 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.944564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.944623 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.944639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.944665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:08 crc kubenswrapper[4914]: I1122 14:28:08.944700 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:08Z","lastTransitionTime":"2025-11-22T14:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.047802 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.047866 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.047911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.047937 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.047958 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.151140 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.151294 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.151313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.151343 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.151361 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.253523 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.253575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.253594 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.253616 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.253631 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.356176 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.356225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.356242 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.356266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.356283 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.459075 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.459130 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.459149 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.459172 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.459190 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.561770 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.561832 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.561854 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.561915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.561938 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.664398 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.664450 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.664467 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.664489 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.664506 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.767414 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.767480 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.767505 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.767529 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.767545 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.867095 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.867156 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:09 crc kubenswrapper[4914]: E1122 14:28:09.867353 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:09 crc kubenswrapper[4914]: E1122 14:28:09.867517 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.871546 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.871633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.871652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.871675 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.871692 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.973542 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.973600 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.973608 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.973621 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:09 crc kubenswrapper[4914]: I1122 14:28:09.973629 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:09Z","lastTransitionTime":"2025-11-22T14:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.076931 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.076974 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.076990 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.077009 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.077025 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.179687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.179716 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.179725 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.179738 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.179746 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.282335 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.282369 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.282377 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.282391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.282403 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.385811 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.385920 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.385941 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.385969 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.385986 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.488815 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.488842 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.488850 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.488863 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.488894 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.591511 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.591577 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.591596 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.591619 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.591635 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.693761 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.693812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.693829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.693850 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.693867 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.796922 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.797082 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.797099 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.797116 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.797131 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.867925 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:10 crc kubenswrapper[4914]: E1122 14:28:10.868110 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.868343 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:10 crc kubenswrapper[4914]: E1122 14:28:10.868440 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.902186 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.902245 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.902262 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.902289 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:10 crc kubenswrapper[4914]: I1122 14:28:10.902319 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:10Z","lastTransitionTime":"2025-11-22T14:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.004762 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.004803 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.004812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.004827 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.004836 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.108361 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.108416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.108433 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.108460 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.108478 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.211495 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.211564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.211588 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.211617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.211637 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.314704 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.314770 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.314787 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.314812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.314833 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.343195 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.343246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.343265 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.343291 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.343310 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.364372 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:11Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.370581 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.370661 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.370680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.370709 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.370729 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.395363 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:11Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.401269 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.401328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.401348 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.401376 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.401397 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.421913 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:11Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.427722 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.427813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.427833 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.427871 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.427939 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.450765 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:11Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.456048 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.456110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.456131 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.456155 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.456174 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.476219 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:11Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.476445 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.478582 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.478633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.478650 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.478672 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.478688 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.582123 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.582164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.582183 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.582204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.582224 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.684962 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.685916 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.686126 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.686327 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.686522 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.789728 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.789801 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.789814 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.789831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.789843 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.867332 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.867349 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.867977 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:11 crc kubenswrapper[4914]: E1122 14:28:11.868233 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.892195 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.892242 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.892279 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.892298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.892311 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.995270 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.995348 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.995372 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.995406 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:11 crc kubenswrapper[4914]: I1122 14:28:11.995429 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:11Z","lastTransitionTime":"2025-11-22T14:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.098215 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.098291 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.098314 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.098340 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.098356 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.201255 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.201634 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.201837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.202044 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.202214 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.304463 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.304749 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.304953 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.305117 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.305239 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.408003 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.408068 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.408083 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.408107 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.408124 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.511482 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.511535 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.511551 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.511573 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.511589 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.614787 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.614859 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.614939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.614971 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.614988 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.718099 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.718167 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.718183 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.718206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.718224 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.821210 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.821279 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.821302 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.821329 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.821349 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.867555 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:12 crc kubenswrapper[4914]: E1122 14:28:12.867689 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.867556 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:12 crc kubenswrapper[4914]: E1122 14:28:12.867955 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.924340 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.924397 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.924414 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.924436 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:12 crc kubenswrapper[4914]: I1122 14:28:12.924454 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:12Z","lastTransitionTime":"2025-11-22T14:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.027758 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.027830 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.027852 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.027918 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.027946 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.131824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.131951 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.131972 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.132014 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.132036 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.235019 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.235081 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.235099 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.235124 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.235147 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.339113 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.339184 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.339201 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.339225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.339246 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.442286 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.442339 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.442357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.442383 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.442401 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.546395 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.546457 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.546477 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.546505 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.546526 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.650419 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.650485 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.650509 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.650540 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.650562 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.754073 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.754266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.754288 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.754336 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.754355 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.858117 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.858198 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.858225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.858255 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.858278 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.867806 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:13 crc kubenswrapper[4914]: E1122 14:28:13.868066 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.868098 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:13 crc kubenswrapper[4914]: E1122 14:28:13.868478 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.961571 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.961639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.961655 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.961681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:13 crc kubenswrapper[4914]: I1122 14:28:13.961700 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:13Z","lastTransitionTime":"2025-11-22T14:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.065041 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.065116 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.065134 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.065161 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.065211 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.168741 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.168813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.168838 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.168869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.168929 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.272532 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.272644 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.272669 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.272700 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.272718 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.375789 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.375911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.375930 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.375952 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.375969 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.479633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.479685 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.479701 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.479723 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.479740 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.582722 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.582758 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.582774 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.582793 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.582810 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.686357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.686430 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.686456 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.686485 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.686506 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.789946 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.790022 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.790045 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.790074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.790096 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.867045 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:14 crc kubenswrapper[4914]: E1122 14:28:14.867313 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.867362 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:14 crc kubenswrapper[4914]: E1122 14:28:14.867500 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.892626 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.896418 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.896632 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.896682 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.896721 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.896749 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.918528 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.937800 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.950808 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.965529 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.980584 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:14 crc kubenswrapper[4914]: I1122 14:28:14.992385 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:14Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:14.999308 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:14.999335 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:14.999343 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:14.999357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:14.999364 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:14Z","lastTransitionTime":"2025-11-22T14:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.013527 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.043816 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.059284 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.078306 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.095418 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.102049 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.102081 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.102111 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.102127 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.102137 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.113491 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.134326 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.168022 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.188605 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.205280 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.205318 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.205330 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.205348 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.205362 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.206833 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.229083 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:15Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.307830 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.307871 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.307896 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.307913 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.307926 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.410367 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.410407 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.410420 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.410438 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.410452 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.512994 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.513031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.513041 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.513056 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.513068 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.616028 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.616065 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.616077 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.616094 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.616108 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.718610 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.718658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.718673 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.718693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.718708 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.821270 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.821324 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.821335 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.821353 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.821364 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.867340 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:15 crc kubenswrapper[4914]: E1122 14:28:15.867542 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.867366 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:15 crc kubenswrapper[4914]: E1122 14:28:15.867769 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.925467 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.925569 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.925593 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.926569 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:15 crc kubenswrapper[4914]: I1122 14:28:15.926986 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:15Z","lastTransitionTime":"2025-11-22T14:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.031131 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.031944 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.032208 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.032393 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.032539 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.135798 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.135863 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.135929 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.135960 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.135976 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.240258 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.240328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.240347 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.240372 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.240390 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.343315 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.343416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.343436 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.343461 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.343479 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.446839 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.446955 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.446974 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.446998 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.447015 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.550586 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.550655 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.550688 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.550830 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.550955 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.653019 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.653081 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.653095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.653110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.653121 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.756120 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.756190 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.756215 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.756241 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.756258 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.860791 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.860855 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.860907 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.860932 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.860950 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.867480 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:16 crc kubenswrapper[4914]: E1122 14:28:16.867648 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.868054 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:16 crc kubenswrapper[4914]: E1122 14:28:16.868194 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.963369 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.963419 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.963435 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.963458 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:16 crc kubenswrapper[4914]: I1122 14:28:16.963474 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:16Z","lastTransitionTime":"2025-11-22T14:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.066601 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.066664 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.066683 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.066710 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.066729 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.171033 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.171091 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.171108 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.171132 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.171154 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.274023 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.274080 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.274093 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.274110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.274121 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.377532 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.377594 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.377609 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.377629 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.377646 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.480611 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.480687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.480708 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.480741 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.480766 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.583759 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.583807 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.583818 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.583835 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.583848 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.685673 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.685725 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.685744 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.685773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.685794 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.788111 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.788171 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.788188 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.788213 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.788229 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.867280 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.867353 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:17 crc kubenswrapper[4914]: E1122 14:28:17.867485 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:17 crc kubenswrapper[4914]: E1122 14:28:17.867659 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.891054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.891246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.891273 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.891307 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.891328 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.993723 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.993761 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.993771 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.993784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:17 crc kubenswrapper[4914]: I1122 14:28:17.993793 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:17Z","lastTransitionTime":"2025-11-22T14:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.096865 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.096948 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.096970 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.096993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.097009 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.200062 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.200117 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.200134 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.200159 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.200179 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.303116 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.303169 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.303181 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.303197 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.303211 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.405403 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.405448 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.405457 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.405471 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.405480 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.508271 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.508311 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.508322 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.508343 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.508354 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.610790 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.610836 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.610847 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.610866 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.610901 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.719202 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.719285 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.719306 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.719331 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.719349 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.822743 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.822812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.822829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.822856 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.822916 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.867407 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.867440 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:18 crc kubenswrapper[4914]: E1122 14:28:18.867595 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:18 crc kubenswrapper[4914]: E1122 14:28:18.867729 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.925744 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.925801 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.925819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.925844 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:18 crc kubenswrapper[4914]: I1122 14:28:18.925862 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:18Z","lastTransitionTime":"2025-11-22T14:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.029160 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.029221 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.029239 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.029264 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.029280 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.131748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.131803 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.131824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.131905 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.131923 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.233970 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.233999 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.234007 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.234021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.234030 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.337244 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.337295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.337306 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.337323 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.337335 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.439850 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.439906 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.439915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.439930 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.439940 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.542479 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.542521 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.542535 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.542551 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.542562 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.645970 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.646052 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.646075 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.646199 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.646227 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.749499 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.749564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.749580 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.749604 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.749621 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.852923 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.852984 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.853001 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.853024 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.853041 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.867778 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.867808 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:19 crc kubenswrapper[4914]: E1122 14:28:19.868869 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:19 crc kubenswrapper[4914]: E1122 14:28:19.869328 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.956167 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.956233 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.956255 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.956305 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:19 crc kubenswrapper[4914]: I1122 14:28:19.956325 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:19Z","lastTransitionTime":"2025-11-22T14:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.058713 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.058745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.058780 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.058796 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.058806 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.161344 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.161398 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.161416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.161440 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.161459 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.264918 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.264987 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.265006 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.265032 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.265050 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.368301 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.368350 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.368366 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.368391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.368408 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.470502 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.470593 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.470629 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.470647 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.470658 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.573126 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.573198 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.573215 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.573239 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.573256 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.676159 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.676220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.676238 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.676262 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.676283 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.779748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.779812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.779829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.779853 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.779870 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.867017 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.867106 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:20 crc kubenswrapper[4914]: E1122 14:28:20.867265 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:20 crc kubenswrapper[4914]: E1122 14:28:20.867377 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.882161 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.882221 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.882246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.882270 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.882295 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.984518 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.984551 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.984560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.984575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:20 crc kubenswrapper[4914]: I1122 14:28:20.984583 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:20Z","lastTransitionTime":"2025-11-22T14:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.086366 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.086404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.086415 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.086431 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.086440 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.188507 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.188540 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.188548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.188562 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.188571 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.291072 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.291110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.291118 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.291131 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.291140 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.393722 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.393772 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.393784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.393804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.393817 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.495801 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.495843 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.495856 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.495887 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.495898 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.598207 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.598239 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.598248 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.598260 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.598269 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.650055 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.650233 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.650298 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:28:53.650279752 +0000 UTC m=+99.458849413 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.700854 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.700909 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.700920 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.700939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.700952 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.763577 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.763623 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.763633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.763651 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.763665 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.782987 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:21Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.789164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.789266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.789289 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.789320 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.789343 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.809227 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:21Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.815335 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.815369 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.815378 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.815396 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.815407 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.828540 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:21Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.833250 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.833432 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.833452 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.833520 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.833545 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.847515 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:21Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.851810 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.851841 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.851851 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.851869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.851930 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.867184 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:21Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.867269 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.867290 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.867294 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.867562 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:21 crc kubenswrapper[4914]: E1122 14:28:21.867685 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.869857 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.869894 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.869906 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.869919 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.869929 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.972405 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.972450 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.972466 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.972488 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:21 crc kubenswrapper[4914]: I1122 14:28:21.972505 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:21Z","lastTransitionTime":"2025-11-22T14:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.074764 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.074798 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.074806 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.074820 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.074830 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.178298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.178383 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.178410 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.178446 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.178474 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.281382 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.281450 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.281461 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.281478 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.281510 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.383994 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.384048 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.384059 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.384076 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.384088 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.487082 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.487169 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.487196 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.487229 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.487253 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.590742 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.590853 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.590915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.590939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.590953 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.694304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.694390 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.694414 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.694447 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.694471 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.797215 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.797309 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.797330 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.797360 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.797380 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.868088 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.868097 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:22 crc kubenswrapper[4914]: E1122 14:28:22.868525 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:22 crc kubenswrapper[4914]: E1122 14:28:22.868689 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.868811 4914 scope.go:117] "RemoveContainer" containerID="6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb" Nov 22 14:28:22 crc kubenswrapper[4914]: E1122 14:28:22.869300 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.900124 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.900164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.900177 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.900196 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:22 crc kubenswrapper[4914]: I1122 14:28:22.900208 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:22Z","lastTransitionTime":"2025-11-22T14:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.002956 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.003029 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.003048 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.003076 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.003100 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.106089 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.106206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.106236 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.106277 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.106300 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.208575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.208627 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.208638 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.208658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.208671 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.312145 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.312206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.312225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.312249 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.312268 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.365414 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/0.log" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.365526 4914 generic.go:334] "Generic (PLEG): container finished" podID="b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed" containerID="6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090" exitCode=1 Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.365592 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerDied","Data":"6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.366486 4914 scope.go:117] "RemoveContainer" containerID="6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.402037 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.415481 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.415513 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.415522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.415536 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.415548 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.418257 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.437454 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.455054 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.477923 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.495792 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.507670 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.518095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.518116 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.518125 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.518151 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.518160 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.522361 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.537438 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.547330 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.557710 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.575156 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.595298 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.607642 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.618851 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.620249 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.620269 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.620278 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.620290 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.620299 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.634378 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.648986 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.659041 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:23Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.723210 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.723240 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.723247 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.723259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.723267 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.825893 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.825930 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.825941 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.825955 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.825964 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.867747 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:23 crc kubenswrapper[4914]: E1122 14:28:23.867994 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.868002 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:23 crc kubenswrapper[4914]: E1122 14:28:23.868305 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.881792 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.928645 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.928699 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.928712 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.928732 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:23 crc kubenswrapper[4914]: I1122 14:28:23.928745 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:23Z","lastTransitionTime":"2025-11-22T14:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.031497 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.031544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.031560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.031585 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.031601 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.134023 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.134308 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.134391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.134477 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.134550 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.238080 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.238386 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.238469 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.238593 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.238678 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.350318 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.350631 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.350699 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.350790 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.350858 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.371405 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/0.log" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.371591 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerStarted","Data":"dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.391010 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.404638 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.418957 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.432589 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.451204 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.453500 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.453549 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.453567 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.453598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.453932 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.470809 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.495334 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.524649 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.544284 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.555925 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.555984 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.556000 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.556019 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.556031 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.562933 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.579961 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.595862 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.616711 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.630633 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.652622 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.658181 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.658271 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.658283 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.658304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.658317 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.666702 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.685527 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.700116 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.714242 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.761578 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.761639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.761653 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.761678 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.761692 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.865716 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.865757 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.865768 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.865788 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.865800 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.867279 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.867338 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:24 crc kubenswrapper[4914]: E1122 14:28:24.867397 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:24 crc kubenswrapper[4914]: E1122 14:28:24.867568 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.886150 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.903416 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.916306 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.927493 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.940092 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.954994 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.965953 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.968206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.968230 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.968240 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.968259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.968271 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:24Z","lastTransitionTime":"2025-11-22T14:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.976969 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:24 crc kubenswrapper[4914]: I1122 14:28:24.989289 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.001804 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:24Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.013680 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.030371 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.045641 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.071623 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.071652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.071660 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.071675 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.071688 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.080841 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.108510 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.140939 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.157648 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.168606 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.174151 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.174177 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.174188 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.174206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.174218 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.181913 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:25Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.276550 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.276582 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.276592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.276607 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.276619 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.378824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.378865 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.378895 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.378914 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.378927 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.481358 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.481391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.481401 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.481418 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.481428 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.583837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.583894 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.583908 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.583923 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.583935 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.686429 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.686465 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.686475 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.686491 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.686502 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.788579 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.788614 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.788624 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.788639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.788652 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.867496 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.867582 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:25 crc kubenswrapper[4914]: E1122 14:28:25.867754 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:25 crc kubenswrapper[4914]: E1122 14:28:25.868055 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.891055 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.891087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.891095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.891107 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.891117 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.996745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.996809 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.996831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.996862 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:25 crc kubenswrapper[4914]: I1122 14:28:25.996914 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:25Z","lastTransitionTime":"2025-11-22T14:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.098854 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.098910 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.098919 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.098938 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.098948 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.201835 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.201946 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.201965 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.201991 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.202013 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.304831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.304871 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.304902 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.304920 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.304932 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.406487 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.406522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.406531 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.406566 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.406576 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.508737 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.508783 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.508794 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.508809 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.508819 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.611893 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.611944 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.611954 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.611971 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.611982 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.714778 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.714819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.714833 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.714852 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.714867 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.817601 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.817664 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.817682 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.817705 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.817723 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.867664 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.867855 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:26 crc kubenswrapper[4914]: E1122 14:28:26.867966 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:26 crc kubenswrapper[4914]: E1122 14:28:26.868153 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.920644 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.920713 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.920733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.920758 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:26 crc kubenswrapper[4914]: I1122 14:28:26.920778 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:26Z","lastTransitionTime":"2025-11-22T14:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.023468 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.023613 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.023631 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.023657 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.023673 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.126690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.126793 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.126811 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.126835 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.126852 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.229368 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.229747 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.230006 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.230302 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.230557 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.333509 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.333552 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.333577 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.333597 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.333609 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.436460 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.436555 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.436581 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.436615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.436639 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.539476 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.539538 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.539554 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.539578 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.539598 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.642398 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.642460 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.642477 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.642501 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.642521 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.745676 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.745739 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.745759 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.745785 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.745803 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.849262 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.849316 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.849331 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.849353 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.849370 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.867407 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:27 crc kubenswrapper[4914]: E1122 14:28:27.867621 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.867933 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:27 crc kubenswrapper[4914]: E1122 14:28:27.868047 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.952552 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.952647 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.952667 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.952695 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:27 crc kubenswrapper[4914]: I1122 14:28:27.952713 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:27Z","lastTransitionTime":"2025-11-22T14:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.054837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.054892 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.054906 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.054923 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.054933 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.157496 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.157540 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.157548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.157563 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.157572 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.260000 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.260043 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.260052 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.260070 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.260080 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.362328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.362369 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.362379 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.362399 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.362412 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.464564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.464609 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.464620 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.464637 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.464646 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.568060 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.568110 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.568123 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.568141 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.568154 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.670756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.670819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.670837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.670863 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.670955 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.774208 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.774277 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.774295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.774332 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.774352 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.867002 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:28 crc kubenswrapper[4914]: E1122 14:28:28.867193 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.867627 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:28 crc kubenswrapper[4914]: E1122 14:28:28.867766 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.876561 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.876609 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.876628 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.876649 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.876669 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.978917 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.978949 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.978959 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.978976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:28 crc kubenswrapper[4914]: I1122 14:28:28.978986 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:28Z","lastTransitionTime":"2025-11-22T14:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.082305 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.082345 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.082357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.082395 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.084210 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.186654 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.186686 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.186695 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.186708 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.186717 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.289733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.289775 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.289784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.289800 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.289811 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.392618 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.392688 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.392705 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.392736 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.392845 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.495662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.495711 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.495729 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.495752 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.495770 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.599058 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.599229 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.599258 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.599349 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.599426 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.702056 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.702106 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.702118 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.702134 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.702145 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.804661 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.804725 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.804737 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.804757 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.804771 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.867738 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.867738 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:29 crc kubenswrapper[4914]: E1122 14:28:29.868031 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:29 crc kubenswrapper[4914]: E1122 14:28:29.868110 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.907071 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.907133 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.907155 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.907183 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:29 crc kubenswrapper[4914]: I1122 14:28:29.907207 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:29Z","lastTransitionTime":"2025-11-22T14:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.010246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.010362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.010439 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.010480 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.010557 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.112622 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.112701 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.112847 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.112949 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.113014 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.215081 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.215109 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.215119 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.215131 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.215143 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.318001 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.318063 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.318084 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.318114 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.318136 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.421046 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.421113 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.421137 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.421164 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.421187 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.524234 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.524301 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.524318 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.524342 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.524359 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.627869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.627964 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.627981 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.628010 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.628034 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.730991 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.731055 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.731074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.731102 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.731119 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.835322 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.835400 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.835424 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.835453 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.835479 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.867196 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.867388 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:30 crc kubenswrapper[4914]: E1122 14:28:30.867606 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:30 crc kubenswrapper[4914]: E1122 14:28:30.867746 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.938522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.938606 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.938626 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.938652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:30 crc kubenswrapper[4914]: I1122 14:28:30.938671 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:30Z","lastTransitionTime":"2025-11-22T14:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.041851 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.041964 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.041990 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.042023 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.042052 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.145541 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.145609 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.145630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.145657 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.145684 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.249775 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.249853 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.249876 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.249964 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.250022 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.353187 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.353244 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.353259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.353281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.353297 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.456274 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.456342 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.456363 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.456390 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.456408 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.559672 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.559727 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.559748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.559775 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.559794 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.662706 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.662781 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.662843 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.662875 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.662935 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.765530 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.765577 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.765591 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.765610 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.765625 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.866790 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.866872 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:31 crc kubenswrapper[4914]: E1122 14:28:31.866982 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:31 crc kubenswrapper[4914]: E1122 14:28:31.867159 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.868688 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.868792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.868813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.868869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.868912 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.971357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.971391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.971402 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.971417 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:31 crc kubenswrapper[4914]: I1122 14:28:31.971428 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:31Z","lastTransitionTime":"2025-11-22T14:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.006449 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.006504 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.006525 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.006557 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.006582 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.031322 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:32Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.036321 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.036404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.036424 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.036453 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.036475 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.057723 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:32Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.063569 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.063646 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.063666 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.063693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.063711 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.085338 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:32Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.089978 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.090024 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.090035 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.090051 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.090062 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.105979 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:32Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.110568 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.110631 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.110648 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.110679 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.110697 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.131254 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:32Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.131479 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.133560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.133612 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.133629 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.133649 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.133665 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.236135 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.236223 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.236252 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.236287 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.236311 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.339663 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.339733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.339751 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.339778 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.339799 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.442305 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.442392 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.442416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.442449 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.442470 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.545812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.545956 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.545989 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.546026 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.546050 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.649534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.649588 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.649599 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.649617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.649631 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.752149 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.752223 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.752234 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.752258 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.752275 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.855897 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.855998 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.856021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.856050 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.856072 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.867202 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.867258 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.867354 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:32 crc kubenswrapper[4914]: E1122 14:28:32.867669 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.959426 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.959468 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.959476 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.959493 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:32 crc kubenswrapper[4914]: I1122 14:28:32.959504 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:32Z","lastTransitionTime":"2025-11-22T14:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.062212 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.062279 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.062301 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.062328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.062346 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.164640 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.164697 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.164714 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.164736 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.164756 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.267413 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.267459 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.267476 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.267499 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.267519 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.370234 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.370298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.370314 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.370340 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.370359 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.473617 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.473703 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.473733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.473774 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.473889 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.576719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.576766 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.576778 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.576798 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.576815 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.679041 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.679080 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.679089 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.679103 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.679115 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.781408 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.781491 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.781513 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.781545 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.781567 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.866889 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.867127 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:33 crc kubenswrapper[4914]: E1122 14:28:33.867173 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:33 crc kubenswrapper[4914]: E1122 14:28:33.867548 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.867920 4914 scope.go:117] "RemoveContainer" containerID="6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.884792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.884824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.884834 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.884849 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.884860 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.988711 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.989195 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.989248 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.989275 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:33 crc kubenswrapper[4914]: I1122 14:28:33.989292 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:33Z","lastTransitionTime":"2025-11-22T14:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.092558 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.092628 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.092649 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.092680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.092700 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.196230 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.196277 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.196295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.196321 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.196339 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.299304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.299359 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.299375 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.299399 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.299415 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.402544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.402597 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.402616 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.402641 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.402661 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.405575 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/2.log" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.410296 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.411239 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.431481 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.453699 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.484527 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.505500 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.507501 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.507532 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.507544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.507565 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.507580 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.522150 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.534790 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.549212 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.561732 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.573616 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.584514 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.594420 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.609639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.609687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.609649 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.609702 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.609918 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.609955 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.624333 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.640776 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.660246 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.682176 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.694123 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.705275 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.713048 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.713098 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.713112 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.713132 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.713146 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.723369 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.815281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.815336 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.815350 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.815373 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.815387 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.867537 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:34 crc kubenswrapper[4914]: E1122 14:28:34.867686 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.867552 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:34 crc kubenswrapper[4914]: E1122 14:28:34.867910 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.880545 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.896921 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.911502 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.917941 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.917976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.917987 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.918004 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.918016 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:34Z","lastTransitionTime":"2025-11-22T14:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.925864 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.940253 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.951805 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.960685 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.970767 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.981113 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:34 crc kubenswrapper[4914]: I1122 14:28:34.992590 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:34Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.005319 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.020415 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.020481 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.020496 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.020516 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.020528 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.022164 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.033549 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.048186 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.063289 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.074128 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.086532 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.106186 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.118350 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.123470 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.123517 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.123527 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.123544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.123555 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.226306 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.226401 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.226420 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.226445 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.226496 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.329780 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.330032 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.330108 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.330174 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.330259 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.415599 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/3.log" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.417261 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/2.log" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.420854 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" exitCode=1 Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.420977 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.421021 4914 scope.go:117] "RemoveContainer" containerID="6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.422689 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:28:35 crc kubenswrapper[4914]: E1122 14:28:35.423171 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.432816 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.432929 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.432953 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.432983 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.433002 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.439754 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.457186 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.474407 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.497061 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a58c576bd50519c9110c5c0796fab508361ca930143458f8ce6f0dd419b38eb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:06Z\\\",\\\"message\\\":\\\"retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939015 6543 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.939020 6543 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1122 14:28:06.939023 6543 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1122 14:28:06.939027 6543 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1122 14:28:06.938677 6543 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-8hpjt\\\\nI1122 14:28:06.939027 6543 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1122 14:28:06.938638 6543 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:34Z\\\",\\\"message\\\":\\\"k-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-fftvp openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb]\\\\nI1122 14:28:34.879039 6890 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1122 14:28:34.879051 6890 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879059 6890 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879065 6890 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb in node crc\\\\nI1122 14:28:34.879070 6890 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb after 0 failed attempt(s)\\\\nI1122 14:28:34.879074 6890 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879087 6890 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:28:34.879136 6890 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.507660 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.520996 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.533093 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.537033 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.537104 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.537129 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.537163 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.537186 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.545321 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.559381 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.591790 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.606548 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.618624 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.632110 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.640304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.640387 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.640403 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.640422 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.640466 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.646915 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.661076 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.672838 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.683429 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.694963 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.712385 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:35Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.742907 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.742951 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.742960 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.742976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.742985 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.845419 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.845463 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.845489 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.845503 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.845512 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.867605 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.867612 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:35 crc kubenswrapper[4914]: E1122 14:28:35.867810 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:35 crc kubenswrapper[4914]: E1122 14:28:35.868106 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.947350 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.947596 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.947687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.947791 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:35 crc kubenswrapper[4914]: I1122 14:28:35.947859 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:35Z","lastTransitionTime":"2025-11-22T14:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.050578 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.050658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.050681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.050710 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.050729 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.153699 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.153757 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.153773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.153795 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.153812 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.258054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.258139 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.258160 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.258189 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.258213 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.360768 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.360851 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.360920 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.360955 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.360979 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.428250 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/3.log" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.433181 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:28:36 crc kubenswrapper[4914]: E1122 14:28:36.433338 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.447814 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.463639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.463715 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.463733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.463760 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.463782 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.464418 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.480200 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.493734 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.505566 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.520680 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.551093 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:34Z\\\",\\\"message\\\":\\\"k-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-fftvp openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb]\\\\nI1122 14:28:34.879039 6890 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1122 14:28:34.879051 6890 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879059 6890 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879065 6890 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb in node crc\\\\nI1122 14:28:34.879070 6890 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb after 0 failed attempt(s)\\\\nI1122 14:28:34.879074 6890 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879087 6890 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:28:34.879136 6890 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.564593 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.567648 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.567708 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.567724 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.567744 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.567759 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.578783 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.591454 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.604444 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.623700 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.643038 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.663618 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.670994 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.671035 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.671078 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.671102 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.671153 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.682094 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.726032 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.748559 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.765691 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.773958 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.773998 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.774009 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.774026 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.774041 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.778215 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:36Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.866944 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.867076 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:36 crc kubenswrapper[4914]: E1122 14:28:36.867251 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:36 crc kubenswrapper[4914]: E1122 14:28:36.867519 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.880290 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.880336 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.880352 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.880372 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.880603 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.984428 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.984534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.984553 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.984615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:36 crc kubenswrapper[4914]: I1122 14:28:36.984636 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:36Z","lastTransitionTime":"2025-11-22T14:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.088550 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.089014 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.089177 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.089327 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.089453 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.193515 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.193589 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.193608 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.193639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.193659 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.297055 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.297121 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.297142 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.297173 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.297194 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.400042 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.400489 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.400754 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.400925 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.401016 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.504479 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.505438 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.505583 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.505731 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.505942 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.609190 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.609267 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.609295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.609328 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.609352 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.712598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.712669 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.712687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.712713 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.712730 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.815806 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.816129 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.816309 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.816498 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.816651 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.867162 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.867212 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:37 crc kubenswrapper[4914]: E1122 14:28:37.867297 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:37 crc kubenswrapper[4914]: E1122 14:28:37.867522 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.919342 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.919375 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.919386 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.919416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:37 crc kubenswrapper[4914]: I1122 14:28:37.919429 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:37Z","lastTransitionTime":"2025-11-22T14:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.022003 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.022283 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.022354 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.022426 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.022494 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.124681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.125000 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.125070 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.125155 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.125246 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.228134 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.228179 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.228188 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.228205 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.228214 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.331579 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.331655 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.331676 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.331705 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.331725 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.435044 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.435117 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.435135 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.435162 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.435180 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.538756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.538844 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.538862 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.538926 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.538977 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.641915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.641971 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.641988 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.642011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.642028 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.745718 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.745780 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.745800 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.745829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.745846 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.748291 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.748404 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.748454 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.748416254 +0000 UTC m=+148.556985956 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.748483 4914 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.748547 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.748529478 +0000 UTC m=+148.557099139 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.748593 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.748682 4914 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.748713 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.748706383 +0000 UTC m=+148.557276044 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849364 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849407 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849467 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849478 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849647 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849491 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849682 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849705 4914 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849738 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.849765 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849789 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.849758082 +0000 UTC m=+148.658327773 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849829 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849976 4914 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.849999 4914 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.850088 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.85006193 +0000 UTC m=+148.658631631 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.867669 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.867685 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.868283 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:38 crc kubenswrapper[4914]: E1122 14:28:38.868432 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.952594 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.952668 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.952688 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.952715 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:38 crc kubenswrapper[4914]: I1122 14:28:38.952739 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:38Z","lastTransitionTime":"2025-11-22T14:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.055831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.055927 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.055945 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.055971 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.055988 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.160999 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.161651 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.161680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.161718 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.161743 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.265576 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.265650 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.265668 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.265695 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.265714 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.368652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.369059 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.369266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.369589 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.369804 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.473609 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.474041 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.474353 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.474517 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.474661 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.577980 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.578054 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.578072 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.578102 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.578122 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.681539 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.681606 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.681625 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.681656 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.681679 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.785562 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.785631 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.785656 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.785689 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.785708 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.867534 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.867631 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:39 crc kubenswrapper[4914]: E1122 14:28:39.867758 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:39 crc kubenswrapper[4914]: E1122 14:28:39.868797 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.888919 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.888980 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.889002 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.889029 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.889048 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.992107 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.992173 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.992191 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.992216 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:39 crc kubenswrapper[4914]: I1122 14:28:39.992235 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:39Z","lastTransitionTime":"2025-11-22T14:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.096748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.096824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.096843 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.096871 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.096928 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.200197 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.200263 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.200282 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.200308 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.200331 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.303537 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.303602 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.303623 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.303649 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.303665 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.407380 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.407462 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.407484 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.407550 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.407573 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.511662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.511753 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.511779 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.511815 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.511839 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.615934 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.616011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.616033 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.616067 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.616091 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.719329 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.719407 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.719459 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.719493 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.719513 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.852706 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.852787 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.852814 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.852845 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.852869 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.867582 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.867861 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:40 crc kubenswrapper[4914]: E1122 14:28:40.868188 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:40 crc kubenswrapper[4914]: E1122 14:28:40.868424 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.956418 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.956564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.956586 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.956648 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:40 crc kubenswrapper[4914]: I1122 14:28:40.956669 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:40Z","lastTransitionTime":"2025-11-22T14:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.060314 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.060385 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.060404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.060429 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.060447 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.164077 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.164176 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.164204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.164247 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.164270 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.268638 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.268707 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.268725 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.268754 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.268773 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.373030 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.373116 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.373137 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.373166 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.373184 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.478442 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.478519 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.478537 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.478568 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.478590 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.582205 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.582264 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.582284 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.582313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.582332 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.685961 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.686055 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.686077 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.686104 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.686124 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.789356 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.789426 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.789442 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.789465 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.789479 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.867260 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.867260 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:41 crc kubenswrapper[4914]: E1122 14:28:41.867587 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:41 crc kubenswrapper[4914]: E1122 14:28:41.867753 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.892236 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.892304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.892318 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.892340 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.892357 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.995592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.995673 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.995693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.995724 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:41 crc kubenswrapper[4914]: I1122 14:28:41.995744 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:41Z","lastTransitionTime":"2025-11-22T14:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.100716 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.100822 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.100842 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.100934 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.100955 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.205362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.205434 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.205451 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.205480 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.205498 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.259091 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.259157 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.259180 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.259214 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.259240 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.282488 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.288738 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.288791 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.288813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.288846 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.288869 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.318603 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.326560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.326635 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.326654 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.326685 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.326705 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.349288 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.355363 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.355424 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.355441 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.355462 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.355479 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.376585 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.382599 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.382847 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.383089 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.383180 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.383214 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.405365 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:42Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.405601 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.408215 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.408269 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.408288 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.408313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.408335 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.512090 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.512160 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.512176 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.512201 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.512218 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.615382 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.615461 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.615489 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.615525 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.615548 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.718905 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.718977 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.718995 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.719021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.719039 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.822461 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.822526 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.822546 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.822576 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.822594 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.867740 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.867859 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.867976 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:42 crc kubenswrapper[4914]: E1122 14:28:42.868121 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.926088 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.926154 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.926174 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.926202 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:42 crc kubenswrapper[4914]: I1122 14:28:42.926224 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:42Z","lastTransitionTime":"2025-11-22T14:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.029812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.029942 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.029971 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.030004 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.030027 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.133479 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.133547 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.133626 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.133658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.133679 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.237642 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.237734 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.237755 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.237786 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.237813 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.341531 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.341612 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.341631 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.341665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.341690 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.444457 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.444540 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.444561 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.444595 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.444615 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.548146 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.548219 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.548238 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.548267 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.548288 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.651746 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.651818 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.651838 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.651869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.651925 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.755859 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.755970 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.755990 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.756023 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.756045 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.859659 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.859735 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.859754 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.859784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.859804 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.867041 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.867074 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:43 crc kubenswrapper[4914]: E1122 14:28:43.867285 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:43 crc kubenswrapper[4914]: E1122 14:28:43.867487 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.963776 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.963849 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.963868 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.963949 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:43 crc kubenswrapper[4914]: I1122 14:28:43.963970 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:43Z","lastTransitionTime":"2025-11-22T14:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.067206 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.067286 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.067304 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.067340 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.067361 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.170526 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.170636 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.170660 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.170693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.170721 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.275560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.275663 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.275689 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.275725 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.275752 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.378812 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.378937 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.378962 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.378993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.379013 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.482227 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.482313 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.482334 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.482368 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.482390 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.585819 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.585923 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.585943 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.585972 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.585994 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.689461 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.689526 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.689545 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.689573 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.689592 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.793316 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.793407 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.793426 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.793465 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.793485 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.867270 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.867426 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:44 crc kubenswrapper[4914]: E1122 14:28:44.867515 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:44 crc kubenswrapper[4914]: E1122 14:28:44.867652 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.893085 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.896492 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.896551 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.896571 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.896595 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.896614 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.919722 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.936246 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.953067 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.968806 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.988200 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:44Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.999641 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:44 crc kubenswrapper[4914]: I1122 14:28:44.999706 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:44.999769 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:44.999954 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.000018 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:44Z","lastTransitionTime":"2025-11-22T14:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.008184 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.025508 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.055713 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.089911 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:34Z\\\",\\\"message\\\":\\\"k-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-fftvp openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb]\\\\nI1122 14:28:34.879039 6890 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1122 14:28:34.879051 6890 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879059 6890 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879065 6890 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb in node crc\\\\nI1122 14:28:34.879070 6890 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb after 0 failed attempt(s)\\\\nI1122 14:28:34.879074 6890 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879087 6890 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:28:34.879136 6890 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.104399 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.104517 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.104544 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.104619 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.104682 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.108372 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.131672 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.153362 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.172659 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.192013 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.207800 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.207869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.207915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.207942 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.207960 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.226195 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.248617 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.264703 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.286513 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:45Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.311479 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.311583 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.311604 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.311640 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.311662 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.415276 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.415356 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.415375 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.415407 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.415428 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.521755 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.521838 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.521863 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.521927 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.522036 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.627116 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.627177 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.627194 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.627222 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.627241 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.730837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.730975 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.730997 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.731025 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.731045 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.834490 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.834582 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.834608 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.834639 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.834659 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.867271 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.867323 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:45 crc kubenswrapper[4914]: E1122 14:28:45.867431 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:45 crc kubenswrapper[4914]: E1122 14:28:45.867653 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.941678 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.941739 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.941752 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.941775 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:45 crc kubenswrapper[4914]: I1122 14:28:45.941789 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:45Z","lastTransitionTime":"2025-11-22T14:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.045545 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.045603 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.045615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.045636 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.045650 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.149458 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.149540 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.149562 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.149592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.149612 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.252601 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.252675 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.252697 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.252724 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.252745 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.355690 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.355733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.355744 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.355759 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.355771 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.459294 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.459378 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.459397 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.459429 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.459449 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.563762 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.563856 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.563915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.563948 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.563972 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.667724 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.667814 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.667834 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.667865 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.667922 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.772422 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.772531 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.772559 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.772600 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.772629 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.867557 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.867612 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:46 crc kubenswrapper[4914]: E1122 14:28:46.867856 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:46 crc kubenswrapper[4914]: E1122 14:28:46.868176 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.875454 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.875526 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.875548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.875572 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.875591 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.979277 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.979321 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.979334 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.979354 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:46 crc kubenswrapper[4914]: I1122 14:28:46.979366 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:46Z","lastTransitionTime":"2025-11-22T14:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.083646 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.083722 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.083739 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.083766 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.083788 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.187083 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.187162 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.187190 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.187222 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.187244 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.291161 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.291234 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.291254 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.291282 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.291304 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.393772 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.393846 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.393861 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.393917 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.393933 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.496402 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.496481 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.496501 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.496530 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.496553 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.601582 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.601680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.601709 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.601745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.601769 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.704104 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.704147 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.704160 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.704177 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.704190 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.807695 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.807748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.807762 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.807782 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.807799 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.867455 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.867473 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:47 crc kubenswrapper[4914]: E1122 14:28:47.867716 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:47 crc kubenswrapper[4914]: E1122 14:28:47.867865 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.911221 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.911317 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.911351 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.911384 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:47 crc kubenswrapper[4914]: I1122 14:28:47.911408 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:47Z","lastTransitionTime":"2025-11-22T14:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.014622 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.014694 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.014717 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.014751 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.014776 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.117757 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.117815 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.117833 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.117857 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.117902 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.222403 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.222482 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.222505 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.222536 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.222566 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.326597 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.326680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.326697 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.326726 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.326748 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.430273 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.430354 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.430373 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.430404 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.430432 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.533516 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.533575 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.533585 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.533604 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.533616 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.636993 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.637082 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.637105 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.637140 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.637163 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.741074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.741138 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.741155 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.741182 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.741201 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.845080 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.845266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.845332 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.845368 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.845389 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.866848 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:48 crc kubenswrapper[4914]: E1122 14:28:48.867278 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.867574 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:48 crc kubenswrapper[4914]: E1122 14:28:48.868025 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.950362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.950520 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.950547 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.950579 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:48 crc kubenswrapper[4914]: I1122 14:28:48.950600 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:48Z","lastTransitionTime":"2025-11-22T14:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.054563 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.054637 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.054840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.054911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.054946 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.159839 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.159931 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.159949 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.159974 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.159992 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.262961 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.263046 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.263065 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.263091 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.263111 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.367100 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.367178 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.367198 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.367616 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.367660 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.471178 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.471228 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.471244 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.471267 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.471284 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.574804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.574915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.574938 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.574973 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.574994 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.677573 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.677636 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.677653 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.677681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.677718 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.781998 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.782069 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.782093 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.782124 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.782144 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.867113 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.867113 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:49 crc kubenswrapper[4914]: E1122 14:28:49.867306 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:49 crc kubenswrapper[4914]: E1122 14:28:49.867469 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.868538 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:28:49 crc kubenswrapper[4914]: E1122 14:28:49.868824 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.885099 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.885150 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.885167 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.885190 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.885209 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.988166 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.988225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.988241 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.988267 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:49 crc kubenswrapper[4914]: I1122 14:28:49.988285 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:49Z","lastTransitionTime":"2025-11-22T14:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.091416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.091499 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.091553 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.091580 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.091599 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.195687 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.195778 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.195805 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.195840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.195864 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.298921 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.298997 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.299025 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.299057 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.299082 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.402372 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.402497 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.402522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.402548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.402565 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.505730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.505808 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.505826 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.505851 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.505871 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.609398 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.609517 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.609537 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.609570 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.609590 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.713332 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.713406 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.713477 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.713508 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.713572 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.817298 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.817403 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.817428 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.817465 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.817492 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.867040 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.867044 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:50 crc kubenswrapper[4914]: E1122 14:28:50.867300 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:50 crc kubenswrapper[4914]: E1122 14:28:50.867669 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.920786 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.920848 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.920867 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.920925 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:50 crc kubenswrapper[4914]: I1122 14:28:50.920947 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:50Z","lastTransitionTime":"2025-11-22T14:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.024705 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.024773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.024792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.024817 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.024837 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.128063 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.128132 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.128150 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.128179 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.128199 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.231969 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.232114 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.232137 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.232169 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.232191 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.335672 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.335756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.335792 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.335823 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.335845 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.439311 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.439378 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.439395 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.439421 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.439455 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.542831 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.542955 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.542976 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.543008 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.543030 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.646564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.646629 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.646646 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.646671 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.646689 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.750410 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.750494 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.750519 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.750552 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.750575 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.853902 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.853975 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.853992 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.854022 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.854044 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.867515 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.867604 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:51 crc kubenswrapper[4914]: E1122 14:28:51.868168 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:51 crc kubenswrapper[4914]: E1122 14:28:51.868265 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.957273 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.957345 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.957361 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.957385 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:51 crc kubenswrapper[4914]: I1122 14:28:51.957406 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:51Z","lastTransitionTime":"2025-11-22T14:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.059935 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.060002 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.060020 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.060045 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.060064 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.163680 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.163751 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.163770 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.163797 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.163817 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.266948 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.267011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.267027 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.267081 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.267098 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.370463 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.370543 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.370565 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.370594 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.370614 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.474253 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.474329 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.474350 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.474380 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.474398 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.544074 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.544151 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.544170 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.544199 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.544219 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.567298 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:52Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.573085 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.573165 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.573183 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.573209 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.573229 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.594976 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:52Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.600140 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.600207 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.600228 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.600259 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.600282 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.628227 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:52Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.633783 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.633869 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.633909 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.633963 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.633981 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.656124 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:52Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.661362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.661449 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.661468 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.661497 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.661546 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.684021 4914 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f45cf250-4ef3-489f-b51a-e18266c6e1d1\\\",\\\"systemUUID\\\":\\\"eefbc55f-3f05-4b13-a0e3-46f39658fddc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:52Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.684386 4914 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.687342 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.687410 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.687434 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.687464 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.687484 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.791230 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.791537 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.791669 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.791996 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.792119 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.867624 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.868097 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.867689 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:52 crc kubenswrapper[4914]: E1122 14:28:52.868462 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.895003 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.895058 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.895079 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.895103 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.895122 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.998307 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.999030 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.999065 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.999091 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:52 crc kubenswrapper[4914]: I1122 14:28:52.999105 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:52Z","lastTransitionTime":"2025-11-22T14:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.103450 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.103522 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.103538 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.103564 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.103582 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.207743 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.207824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.207844 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.207900 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.207928 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.311615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.311685 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.311704 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.311732 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.311755 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.414966 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.415071 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.415102 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.415140 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.415167 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.518657 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.518710 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.518723 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.518746 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.518763 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.621305 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.621370 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.621393 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.621420 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.621439 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.725267 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.725346 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.725366 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.725395 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.725414 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.736408 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:53 crc kubenswrapper[4914]: E1122 14:28:53.736686 4914 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:28:53 crc kubenswrapper[4914]: E1122 14:28:53.736865 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs podName:d8b0b491-074a-465a-9419-6481d7877ecf nodeName:}" failed. No retries permitted until 2025-11-22 14:29:57.736817096 +0000 UTC m=+163.545386947 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs") pod "network-metrics-daemon-47bbd" (UID: "d8b0b491-074a-465a-9419-6481d7877ecf") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.830766 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.830836 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.830855 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.830922 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.830943 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.867689 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.867702 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:53 crc kubenswrapper[4914]: E1122 14:28:53.867923 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:53 crc kubenswrapper[4914]: E1122 14:28:53.868124 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.934548 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.934652 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.934677 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.934702 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:53 crc kubenswrapper[4914]: I1122 14:28:53.934724 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:53Z","lastTransitionTime":"2025-11-22T14:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.038730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.038801 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.038824 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.038853 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.038906 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.142362 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.142442 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.142473 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.142506 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.142530 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.245852 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.245969 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.246020 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.246050 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.246069 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.349513 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.349612 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.349625 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.349641 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.349654 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.452549 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.452615 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.452633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.452658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.452676 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.576353 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.576454 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.576481 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.576515 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.576551 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.680651 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.680728 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.680748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.680784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.680805 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.784165 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.784235 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.784254 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.784281 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.784300 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.867695 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.867870 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:54 crc kubenswrapper[4914]: E1122 14:28:54.868135 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:54 crc kubenswrapper[4914]: E1122 14:28:54.868365 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.885171 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c6036f-4eae-46bf-a70f-aa6d73e8f252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25729c086b5adc9564a9b84fd7b0fb138ac6aed7d331e3b3770718938c4d4814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ca2e73afe55d059403605ba2bad98cfe5bef346e55455112d20732b42c60d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.886941 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.886995 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.887013 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.887087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.887109 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.907237 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faeff856ac6ce9c0c5a71f86cb4428e0f17830ae61d67cf2b1d32cf502b50afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50c07e95ef42cc464b59b85e2b703997594cc046fd95330ddb88acb04061c19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.927140 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b64a09-b0a3-4776-93f7-01c4ea3e1388\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e1cf050bd2a37312ca86740848b5661c052b7683b8a09bf11342c30d07fad23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5skzk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.943057 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zpm9b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e6af9df-00d9-4baf-b5a5-dd22b5cc138e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ae88324f38bfc8021bbf965c294ea02ec6461e55246e85f4e88f6b295dee58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lz22n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:37Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zpm9b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.967063 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8eb63cc-74ad-43b4-9bf9-2cac458d53aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c86e51ca55e32266ea1aeca30337b28d671a41e766981b826f827d84e32ea90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fcea5835556d826116d31a0728dc1515b26746cf5aee95149b05a0fc1757c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb10a57d83284c0c6445d2072f599695c4967efb9b3e6c13da74d2cb0550a1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a438cbed2128a5a078e49219f421d5476e229396a21c0e959222067aa2a0260\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://474b670eb0cb69364c3dce8690c3f0f75407d5d6dd496d4aa18b536b5da64f6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d00f69dc9d71dec796705aab437913b8cc35a29e540b305f46d25fac3264ce5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2256995dd9a17bec9830251952a9ad89941eb900c39ff7b461275adebfac0992\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4f765\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jz8w8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.989221 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.989292 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.989310 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.989338 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:54 crc kubenswrapper[4914]: I1122 14:28:54.989358 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:54Z","lastTransitionTime":"2025-11-22T14:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:54.999851 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41ea1017-213b-4be4-87ae-d68655d58a12\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:34Z\\\",\\\"message\\\":\\\"k-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-fftvp openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb]\\\\nI1122 14:28:34.879039 6890 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1122 14:28:34.879051 6890 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879059 6890 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879065 6890 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb in node crc\\\\nI1122 14:28:34.879070 6890 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb after 0 failed attempt(s)\\\\nI1122 14:28:34.879074 6890 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb\\\\nI1122 14:28:34.879087 6890 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1122 14:28:34.879136 6890 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:28:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6j94x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fftvp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:54Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.018474 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-47bbd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8b0b491-074a-465a-9419-6481d7877ecf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjz48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:49Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-47bbd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.039404 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"683b1dec-b37f-4f16-b122-7706bb213e19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30329a0314d477ca57d2ae72d5a157852d005505e45c0f7a49f8ed2644b4fd0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e2ea052ae8f28c90cd0af734d8cbca25bccc9962381b4e9157837e79806bd24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dff1da7541128bc942b2b8da6f07527b2906fb4e490a7bd63dda38eec0f4921\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d9ed3b15a13aa630fae3e6a4fcf1f1f289836378c2b1c6a7aaa4436ac9f996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.059385 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.077587 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2c9e072d911bbcb37e9c76dc08f9355bdd5001d7cdf87c947ad616a4d36790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.093133 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.093567 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.093739 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.093968 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.094177 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.097304 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.130216 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455d76cc-5f07-401a-ad0e-50bcee4b8a65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9af4de92828261e7a0208f4c37ca33e6e192a8576fb94d4def213469355fe2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe8943713c1b882cdda77b307a354917114615b19edd3f6f781dca51eab91fc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52e57a30e907278f84db9c179fe6bc692804a6b21f7b1f02042c1b555df77e9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37d37cbda211bf9f6ace5b9950f3e36a5aa648c4783040e2667e3a146b132913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4686a24c5b0a8cc5d02bff8cf4825a6c73a2a54fcf4473c477014a872bb0a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://803a35b039bdd9295842016ed16a4eb0d5c4d40b7176bce9b193303b2c06fb17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8990e900935f0252c5396e16f3f08c4a55d1300b0be6c99a05215f23bf49c502\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe15b159fa550ec8bc216f13299228b0a1f4b069484aa0eaeff0af1bd9c34d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.149512 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb33e13a-15c5-44fa-a214-c25a60a75e32\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa26a7bae02d360014be01cfc0574fc052985a7060377400508c442a5eeb1f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab4c03da26948b7781afd1953a92688cba21e4bcdc05b80198ad9398d7d00a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd88454d7823f4751953009e6aba45f5491faebad9146838efbc8ecf91d3547\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://156c338134d61b422aeade38eaa8de2abd30ae3603d11821641827f9ed1184ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f50476f1dd718349be119c5307b045c53bd509bea49c0cc5772f7b5795ec8ca1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1122 14:27:19.191019 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1122 14:27:19.191773 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3987253528/tls.crt::/tmp/serving-cert-3987253528/tls.key\\\\\\\"\\\\nI1122 14:27:34.330149 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1122 14:27:34.335450 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1122 14:27:34.335485 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1122 14:27:34.335523 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1122 14:27:34.335533 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1122 14:27:34.348416 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1122 14:27:34.348461 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1122 14:27:34.348481 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1122 14:27:34.348488 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nI1122 14:27:34.348490 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1122 14:27:34.348494 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1122 14:27:34.348570 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1122 14:27:34.352312 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8a58c2392ea1ad710914f58bbf2f2bd6b6b84bbf3b05f05f1536ba313d1d874\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc32b94cb750bed0ed71bb945551b3ab626a37a486b3023dc976926f7a5d4f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.168110 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af8aa724-2af4-47a0-9284-51c3507f2c54\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f6af2582d23d76b32b3003e9f2595163d471247094af406f7aefec2e0199e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e52e9d788e4eeb867f64967bc628c594e48ef4a230f9b30229b2915649821282\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8600327c2e2cdf4ae44e95c125840c76152216788518de03deb6ce53336c23af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fef7b595f766aaac2dfc14405bdefdce7f44eb6d566016c84bd6cda4f95fa7c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-22T14:27:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.186818 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.197630 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.197683 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.197699 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.197721 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.197736 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.204226 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caed1ee765889a13352e1e503dfab173601c35e917b8d606ccc81ee590f94122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.223756 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8hpjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:28:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-22T14:28:23Z\\\",\\\"message\\\":\\\"2025-11-22T14:27:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9\\\\n2025-11-22T14:27:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ef09b35b-646b-48b2-a357-a7e92a9b80c9 to /host/opt/cni/bin/\\\\n2025-11-22T14:27:38Z [verbose] multus-daemon started\\\\n2025-11-22T14:27:38Z [verbose] Readiness Indicator file check\\\\n2025-11-22T14:28:23Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-22T14:27:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:28:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z46pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8hpjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.239228 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f4twn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a5e8620-8854-4329-8935-a984f516d6b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e098fdd8e5b87328c92fcdda1692703b152182329929a770c64c77e830481f27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98f7g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:35Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f4twn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.256066 4914 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be05ffa4-a006-46ca-996e-f5f4c4e0c42a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-22T14:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fffb90c2a3190603797df7cf5e8bf5fdc993178b098c03cc20b512be707705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f56ff65507697c05dfae44c738d88bacaf2029bc1427ddeb3d84d53fffd80206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-22T14:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ms2wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-22T14:27:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mdlfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-22T14:28:55Z is after 2025-08-24T17:21:41Z" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.301091 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.301163 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.301176 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.301196 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.301214 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.404497 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.404610 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.404633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.404694 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.404714 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.507826 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.507911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.507929 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.507950 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.507969 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.610478 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.610558 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.610577 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.610604 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.610623 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.713411 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.713485 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.713501 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.713527 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.713546 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.817732 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.817807 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.817826 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.817854 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.817899 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.866939 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.866973 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:55 crc kubenswrapper[4914]: E1122 14:28:55.867175 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:55 crc kubenswrapper[4914]: E1122 14:28:55.867318 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.921753 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.921814 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.921837 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.921870 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:55 crc kubenswrapper[4914]: I1122 14:28:55.921933 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:55Z","lastTransitionTime":"2025-11-22T14:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.024640 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.024745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.024771 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.024806 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.024832 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.127821 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.127909 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.127928 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.127952 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.127969 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.231593 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.231661 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.231679 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.231705 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.231725 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.335262 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.335335 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.335358 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.335387 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.335405 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.439159 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.439246 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.439265 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.439289 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.439307 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.542720 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.542773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.542795 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.542826 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.542850 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.646407 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.646470 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.646491 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.646523 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.646544 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.749811 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.749899 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.749920 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.749981 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.750003 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.853458 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.853534 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.853560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.853598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.853623 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.868250 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.868348 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:56 crc kubenswrapper[4914]: E1122 14:28:56.868677 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:56 crc kubenswrapper[4914]: E1122 14:28:56.868946 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.956802 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.956860 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.956911 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.956943 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:56 crc kubenswrapper[4914]: I1122 14:28:56.956964 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:56Z","lastTransitionTime":"2025-11-22T14:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.061204 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.061592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.061613 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.061643 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.061662 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.164662 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.164724 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.164741 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.164768 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.164787 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.268498 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.268560 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.268578 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.268602 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.268619 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.372491 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.372552 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.372569 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.372598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.372616 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.475596 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.475660 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.475674 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.475693 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.475706 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.578689 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.578762 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.578773 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.578788 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.578816 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.682319 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.682374 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.682391 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.682416 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.682439 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.790945 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.791021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.791132 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.791162 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.791181 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.867936 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.867936 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:57 crc kubenswrapper[4914]: E1122 14:28:57.868252 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:57 crc kubenswrapper[4914]: E1122 14:28:57.868330 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.894266 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.894314 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.894323 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.894349 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.894359 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.997969 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.998064 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.998077 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.998095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:57 crc kubenswrapper[4914]: I1122 14:28:57.998110 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:57Z","lastTransitionTime":"2025-11-22T14:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.101730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.101796 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.101813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.101840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.101857 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.205051 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.205119 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.205141 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.205167 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.205188 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.307732 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.307769 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.307780 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.307796 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.307809 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.410702 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.410738 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.410748 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.410763 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.410773 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.513730 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.513859 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.513968 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.514050 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.514081 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.617479 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.617556 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.617577 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.617609 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.617628 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.720633 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.720719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.720732 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.720750 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.720763 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.823957 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.824031 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.824050 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.824077 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.824095 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.867694 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.867813 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:28:58 crc kubenswrapper[4914]: E1122 14:28:58.867909 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:28:58 crc kubenswrapper[4914]: E1122 14:28:58.868085 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.927079 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.927144 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.927161 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.927191 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:58 crc kubenswrapper[4914]: I1122 14:28:58.927209 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:58Z","lastTransitionTime":"2025-11-22T14:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.030006 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.030083 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.030102 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.030132 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.030155 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.133366 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.133435 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.133456 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.133481 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.133499 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.236586 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.236665 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.236682 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.236709 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.236728 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.339710 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.339862 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.339927 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.339956 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.339977 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.443484 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.443598 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.443654 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.443681 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.443699 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.546653 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.546756 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.546776 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.546804 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.546826 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.650527 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.650599 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.650619 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.650647 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.650664 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.753719 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.753785 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.753807 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.753840 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.753861 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.856973 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.857064 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.857087 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.857113 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.857135 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.867793 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.867829 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:28:59 crc kubenswrapper[4914]: E1122 14:28:59.868048 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:28:59 crc kubenswrapper[4914]: E1122 14:28:59.868194 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.961218 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.961323 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.961344 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.961373 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:28:59 crc kubenswrapper[4914]: I1122 14:28:59.961396 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:28:59Z","lastTransitionTime":"2025-11-22T14:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.064082 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.064137 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.064150 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.064166 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.064176 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.167294 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.167352 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.167371 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.167388 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.167400 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.269829 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.269888 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.269897 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.269912 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.269922 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.373357 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.373442 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.373463 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.373488 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.373507 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.476918 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.477075 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.477096 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.477121 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.477146 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.580523 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.580589 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.580608 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.580635 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.580653 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.683225 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.683307 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.683331 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.683365 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.683385 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.786986 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.787060 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.787084 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.787114 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.787133 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.867713 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.867777 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:00 crc kubenswrapper[4914]: E1122 14:29:00.867925 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:00 crc kubenswrapper[4914]: E1122 14:29:00.868128 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.890034 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.890178 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.890211 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.890244 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.890266 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.994011 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.994057 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.994073 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.994095 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:00 crc kubenswrapper[4914]: I1122 14:29:00.994112 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:00Z","lastTransitionTime":"2025-11-22T14:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.096682 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.096745 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.096767 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.096795 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.096818 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.200621 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.200707 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.200733 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.200767 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.200793 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.304846 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.304978 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.304999 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.305024 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.305043 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.408546 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.408635 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.408658 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.408684 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.408701 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.511926 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.511996 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.512015 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.512042 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.512066 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.615932 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.616003 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.616021 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.616052 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.616076 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.719161 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.719238 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.719269 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.719295 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.719313 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.823217 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.823292 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.823308 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.823336 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.823355 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.867271 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.867290 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:01 crc kubenswrapper[4914]: E1122 14:29:01.867761 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:01 crc kubenswrapper[4914]: E1122 14:29:01.868034 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.869120 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:29:01 crc kubenswrapper[4914]: E1122 14:29:01.869418 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fftvp_openshift-ovn-kubernetes(41ea1017-213b-4be4-87ae-d68655d58a12)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.926784 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.926832 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.926849 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.926872 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:01 crc kubenswrapper[4914]: I1122 14:29:01.926928 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:01Z","lastTransitionTime":"2025-11-22T14:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.030080 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.030177 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.030194 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.030220 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.030238 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.132844 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.132958 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.132983 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.133012 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.133034 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.236922 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.236995 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.237015 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.237042 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.237062 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.340974 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.341024 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.341036 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.341057 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.341076 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.444841 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.444939 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.444956 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.444983 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.445002 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.552958 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.553008 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.553026 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.553050 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.553063 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.655134 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.655173 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.655186 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.655203 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.655215 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.756813 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.756859 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.756891 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.756915 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.756928 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.801545 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.801583 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.801592 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.801606 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.801616 4914 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-22T14:29:02Z","lastTransitionTime":"2025-11-22T14:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.858831 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2"] Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.859305 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.861447 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.861533 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.862287 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.862358 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.866701 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.866784 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:02 crc kubenswrapper[4914]: E1122 14:29:02.866828 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:02 crc kubenswrapper[4914]: E1122 14:29:02.866957 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.877445 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podStartSLOduration=87.877418184 podStartE2EDuration="1m27.877418184s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:02.877057214 +0000 UTC m=+108.685626885" watchObservedRunningTime="2025-11-22 14:29:02.877418184 +0000 UTC m=+108.685987875" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.903552 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=39.903530532 podStartE2EDuration="39.903530532s" podCreationTimestamp="2025-11-22 14:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:02.903061341 +0000 UTC m=+108.711631042" watchObservedRunningTime="2025-11-22 14:29:02.903530532 +0000 UTC m=+108.712100193" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.903773 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zpm9b" podStartSLOduration=87.903767908 podStartE2EDuration="1m27.903767908s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:02.891417472 +0000 UTC m=+108.699987143" watchObservedRunningTime="2025-11-22 14:29:02.903767908 +0000 UTC m=+108.712337569" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.954869 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80be6282-0275-4e2b-ba69-30309fa80162-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.954950 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80be6282-0275-4e2b-ba69-30309fa80162-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.954998 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/80be6282-0275-4e2b-ba69-30309fa80162-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.955039 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/80be6282-0275-4e2b-ba69-30309fa80162-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.955067 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80be6282-0275-4e2b-ba69-30309fa80162-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:02 crc kubenswrapper[4914]: I1122 14:29:02.980239 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jz8w8" podStartSLOduration=87.980211038 podStartE2EDuration="1m27.980211038s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:02.979866009 +0000 UTC m=+108.788435660" watchObservedRunningTime="2025-11-22 14:29:02.980211038 +0000 UTC m=+108.788780689" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056514 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80be6282-0275-4e2b-ba69-30309fa80162-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056584 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/80be6282-0275-4e2b-ba69-30309fa80162-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056670 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80be6282-0275-4e2b-ba69-30309fa80162-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056728 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80be6282-0275-4e2b-ba69-30309fa80162-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056781 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/80be6282-0275-4e2b-ba69-30309fa80162-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056796 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/80be6282-0275-4e2b-ba69-30309fa80162-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.056899 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/80be6282-0275-4e2b-ba69-30309fa80162-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.058074 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80be6282-0275-4e2b-ba69-30309fa80162-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.062938 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80be6282-0275-4e2b-ba69-30309fa80162-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.078232 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.078190479 podStartE2EDuration="1m28.078190479s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.037040555 +0000 UTC m=+108.845610226" watchObservedRunningTime="2025-11-22 14:29:03.078190479 +0000 UTC m=+108.886760150" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.101859 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80be6282-0275-4e2b-ba69-30309fa80162-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fj8g2\" (UID: \"80be6282-0275-4e2b-ba69-30309fa80162\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.125949 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.125918862 podStartE2EDuration="59.125918862s" podCreationTimestamp="2025-11-22 14:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.104222776 +0000 UTC m=+108.912792457" watchObservedRunningTime="2025-11-22 14:29:03.125918862 +0000 UTC m=+108.934488523" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.160858 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=87.160833197 podStartE2EDuration="1m27.160833197s" podCreationTimestamp="2025-11-22 14:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.159325718 +0000 UTC m=+108.967895389" watchObservedRunningTime="2025-11-22 14:29:03.160833197 +0000 UTC m=+108.969402868" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.175430 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.17541288 podStartE2EDuration="1m28.17541288s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.173549333 +0000 UTC m=+108.982119004" watchObservedRunningTime="2025-11-22 14:29:03.17541288 +0000 UTC m=+108.983982541" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.177148 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.188055 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-f4twn" podStartSLOduration=88.188025964 podStartE2EDuration="1m28.188025964s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.187352597 +0000 UTC m=+108.995922258" watchObservedRunningTime="2025-11-22 14:29:03.188025964 +0000 UTC m=+108.996595635" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.201326 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mdlfb" podStartSLOduration=88.201305794 podStartE2EDuration="1m28.201305794s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.200214486 +0000 UTC m=+109.008784167" watchObservedRunningTime="2025-11-22 14:29:03.201305794 +0000 UTC m=+109.009875465" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.232888 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8hpjt" podStartSLOduration=88.232846323 podStartE2EDuration="1m28.232846323s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.232500533 +0000 UTC m=+109.041070204" watchObservedRunningTime="2025-11-22 14:29:03.232846323 +0000 UTC m=+109.041415994" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.539524 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" event={"ID":"80be6282-0275-4e2b-ba69-30309fa80162","Type":"ContainerStarted","Data":"5ba49115960fc13118fe878be45228ba7bb7e0b2770db9c5bbf78db8bf103252"} Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.539626 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" event={"ID":"80be6282-0275-4e2b-ba69-30309fa80162","Type":"ContainerStarted","Data":"67eb0cdf1fa1a4d1b5c80df8fa9582e1955aac8033a1f7bd98accc4b08b7dc8b"} Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.567209 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fj8g2" podStartSLOduration=88.567188151 podStartE2EDuration="1m28.567188151s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:03.566562466 +0000 UTC m=+109.375132167" watchObservedRunningTime="2025-11-22 14:29:03.567188151 +0000 UTC m=+109.375757822" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.867237 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:03 crc kubenswrapper[4914]: I1122 14:29:03.867349 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:03 crc kubenswrapper[4914]: E1122 14:29:03.867519 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:03 crc kubenswrapper[4914]: E1122 14:29:03.867682 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:04 crc kubenswrapper[4914]: I1122 14:29:04.867079 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:04 crc kubenswrapper[4914]: I1122 14:29:04.867156 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:04 crc kubenswrapper[4914]: E1122 14:29:04.869045 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:04 crc kubenswrapper[4914]: E1122 14:29:04.869142 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:05 crc kubenswrapper[4914]: I1122 14:29:05.867608 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:05 crc kubenswrapper[4914]: I1122 14:29:05.867670 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:05 crc kubenswrapper[4914]: E1122 14:29:05.867750 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:05 crc kubenswrapper[4914]: E1122 14:29:05.867998 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:06 crc kubenswrapper[4914]: I1122 14:29:06.867556 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:06 crc kubenswrapper[4914]: I1122 14:29:06.867560 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:06 crc kubenswrapper[4914]: E1122 14:29:06.867765 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:06 crc kubenswrapper[4914]: E1122 14:29:06.867899 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:07 crc kubenswrapper[4914]: I1122 14:29:07.866867 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:07 crc kubenswrapper[4914]: I1122 14:29:07.866933 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:07 crc kubenswrapper[4914]: E1122 14:29:07.867032 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:07 crc kubenswrapper[4914]: E1122 14:29:07.867258 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:08 crc kubenswrapper[4914]: I1122 14:29:08.867784 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:08 crc kubenswrapper[4914]: E1122 14:29:08.868286 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:08 crc kubenswrapper[4914]: I1122 14:29:08.867787 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:08 crc kubenswrapper[4914]: E1122 14:29:08.868674 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.558196 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/1.log" Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.558851 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/0.log" Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.558979 4914 generic.go:334] "Generic (PLEG): container finished" podID="b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed" containerID="dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879" exitCode=1 Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.559034 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerDied","Data":"dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879"} Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.559106 4914 scope.go:117] "RemoveContainer" containerID="6f87e84be39f8049ebf78e12ec93051d928fd7402e26c4d9a0b5c32c1dff9090" Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.559594 4914 scope.go:117] "RemoveContainer" containerID="dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879" Nov 22 14:29:09 crc kubenswrapper[4914]: E1122 14:29:09.559866 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8hpjt_openshift-multus(b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed)\"" pod="openshift-multus/multus-8hpjt" podUID="b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed" Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.867830 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:09 crc kubenswrapper[4914]: I1122 14:29:09.867970 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:09 crc kubenswrapper[4914]: E1122 14:29:09.868127 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:09 crc kubenswrapper[4914]: E1122 14:29:09.868230 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:10 crc kubenswrapper[4914]: I1122 14:29:10.564607 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/1.log" Nov 22 14:29:10 crc kubenswrapper[4914]: I1122 14:29:10.867139 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:10 crc kubenswrapper[4914]: I1122 14:29:10.867139 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:10 crc kubenswrapper[4914]: E1122 14:29:10.867333 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:10 crc kubenswrapper[4914]: E1122 14:29:10.867459 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:11 crc kubenswrapper[4914]: I1122 14:29:11.867684 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:11 crc kubenswrapper[4914]: I1122 14:29:11.867774 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:11 crc kubenswrapper[4914]: E1122 14:29:11.867831 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:11 crc kubenswrapper[4914]: E1122 14:29:11.868175 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:12 crc kubenswrapper[4914]: I1122 14:29:12.866722 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:12 crc kubenswrapper[4914]: E1122 14:29:12.866856 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:12 crc kubenswrapper[4914]: I1122 14:29:12.866992 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:12 crc kubenswrapper[4914]: E1122 14:29:12.867216 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:13 crc kubenswrapper[4914]: I1122 14:29:13.867331 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:13 crc kubenswrapper[4914]: I1122 14:29:13.867323 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:13 crc kubenswrapper[4914]: E1122 14:29:13.867546 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:13 crc kubenswrapper[4914]: E1122 14:29:13.867683 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:14 crc kubenswrapper[4914]: E1122 14:29:14.825061 4914 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 22 14:29:14 crc kubenswrapper[4914]: I1122 14:29:14.867491 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:14 crc kubenswrapper[4914]: I1122 14:29:14.867606 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:14 crc kubenswrapper[4914]: E1122 14:29:14.869521 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:14 crc kubenswrapper[4914]: E1122 14:29:14.869801 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:14 crc kubenswrapper[4914]: E1122 14:29:14.980181 4914 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 14:29:15 crc kubenswrapper[4914]: I1122 14:29:15.866847 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:15 crc kubenswrapper[4914]: I1122 14:29:15.866985 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:15 crc kubenswrapper[4914]: E1122 14:29:15.867116 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:15 crc kubenswrapper[4914]: E1122 14:29:15.867365 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:15 crc kubenswrapper[4914]: I1122 14:29:15.869259 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.587776 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/3.log" Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.599920 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerStarted","Data":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.601157 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.646446 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podStartSLOduration=101.646421273 podStartE2EDuration="1m41.646421273s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:16.645731385 +0000 UTC m=+122.454301046" watchObservedRunningTime="2025-11-22 14:29:16.646421273 +0000 UTC m=+122.454990944" Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.867279 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.867282 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:16 crc kubenswrapper[4914]: E1122 14:29:16.867678 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:16 crc kubenswrapper[4914]: E1122 14:29:16.867919 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.932600 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-47bbd"] Nov 22 14:29:16 crc kubenswrapper[4914]: I1122 14:29:16.932741 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:16 crc kubenswrapper[4914]: E1122 14:29:16.932863 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:17 crc kubenswrapper[4914]: I1122 14:29:17.867691 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:17 crc kubenswrapper[4914]: E1122 14:29:17.868107 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:18 crc kubenswrapper[4914]: I1122 14:29:18.867328 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:18 crc kubenswrapper[4914]: I1122 14:29:18.867361 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:18 crc kubenswrapper[4914]: E1122 14:29:18.867515 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:18 crc kubenswrapper[4914]: I1122 14:29:18.867749 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:18 crc kubenswrapper[4914]: E1122 14:29:18.867820 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:18 crc kubenswrapper[4914]: E1122 14:29:18.868003 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:19 crc kubenswrapper[4914]: I1122 14:29:19.867282 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:19 crc kubenswrapper[4914]: E1122 14:29:19.867494 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:19 crc kubenswrapper[4914]: E1122 14:29:19.981562 4914 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 14:29:20 crc kubenswrapper[4914]: I1122 14:29:20.867496 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:20 crc kubenswrapper[4914]: I1122 14:29:20.867675 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:20 crc kubenswrapper[4914]: I1122 14:29:20.867729 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:20 crc kubenswrapper[4914]: I1122 14:29:20.868060 4914 scope.go:117] "RemoveContainer" containerID="dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879" Nov 22 14:29:20 crc kubenswrapper[4914]: E1122 14:29:20.868236 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:20 crc kubenswrapper[4914]: E1122 14:29:20.868385 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:20 crc kubenswrapper[4914]: E1122 14:29:20.868533 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:21 crc kubenswrapper[4914]: I1122 14:29:21.620427 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/1.log" Nov 22 14:29:21 crc kubenswrapper[4914]: I1122 14:29:21.620857 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerStarted","Data":"bdb8ada03e70d054a4dfb6fa8af17dd17837b11a178753fb893b8fa6403e4862"} Nov 22 14:29:21 crc kubenswrapper[4914]: I1122 14:29:21.867769 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:21 crc kubenswrapper[4914]: E1122 14:29:21.868015 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:22 crc kubenswrapper[4914]: I1122 14:29:22.867214 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:22 crc kubenswrapper[4914]: I1122 14:29:22.867291 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:22 crc kubenswrapper[4914]: I1122 14:29:22.867214 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:22 crc kubenswrapper[4914]: E1122 14:29:22.867427 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:22 crc kubenswrapper[4914]: E1122 14:29:22.867640 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:22 crc kubenswrapper[4914]: E1122 14:29:22.867837 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:23 crc kubenswrapper[4914]: I1122 14:29:23.867474 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:23 crc kubenswrapper[4914]: E1122 14:29:23.867649 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 22 14:29:24 crc kubenswrapper[4914]: I1122 14:29:24.867643 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:24 crc kubenswrapper[4914]: I1122 14:29:24.867660 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:24 crc kubenswrapper[4914]: E1122 14:29:24.869771 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 22 14:29:24 crc kubenswrapper[4914]: I1122 14:29:24.869867 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:24 crc kubenswrapper[4914]: E1122 14:29:24.870049 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 22 14:29:24 crc kubenswrapper[4914]: E1122 14:29:24.870324 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-47bbd" podUID="d8b0b491-074a-465a-9419-6481d7877ecf" Nov 22 14:29:25 crc kubenswrapper[4914]: I1122 14:29:25.866836 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:25 crc kubenswrapper[4914]: I1122 14:29:25.870407 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 22 14:29:25 crc kubenswrapper[4914]: I1122 14:29:25.871156 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.867095 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.867167 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.867116 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.871267 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.871505 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.874379 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 22 14:29:26 crc kubenswrapper[4914]: I1122 14:29:26.876142 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.587424 4914 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.629672 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vhs7k"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.630731 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.632482 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-gsnrn"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.633095 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.634847 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.635383 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.637376 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.637617 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.638317 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.638567 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.638763 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.639056 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.639189 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.639235 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.640779 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.640845 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.641033 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.641143 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.641795 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vb7pw"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.648328 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.650860 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.653861 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.654388 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.667304 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.667636 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.667966 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.668061 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.668293 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.668488 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.668658 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.668795 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.668950 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669089 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669258 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669281 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669288 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669091 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669393 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669486 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669524 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669558 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.669646 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.670251 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nctzt"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.670734 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.670784 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.671770 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-l6mzq"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.672367 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.672730 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.677356 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-wshgn"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.678046 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.678046 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.681357 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.681820 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kwffm"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.682146 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-w6692"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.682526 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.683288 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.683847 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.696403 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.697084 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.698290 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vhs7k"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.701056 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.701525 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.701672 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.703308 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.703871 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mmm9"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.704463 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nncgs"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.704513 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.705497 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.705712 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.706849 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.711083 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.711728 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.712243 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.712643 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.721166 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-447c5"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.721716 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.722101 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.723018 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.723547 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.723761 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.729677 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pmg89"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.731678 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.735308 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-xfwmr"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.735986 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.761562 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.763810 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.765098 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.765413 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.765622 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.765819 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.766037 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.766249 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.766436 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.766625 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.766853 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.767909 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.768335 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.768549 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.768742 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.773364 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.773644 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.773904 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.774191 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.774389 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.774780 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.775223 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.775319 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.774922 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.775803 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.775816 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.775868 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.775939 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.776051 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.777689 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.792633 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.792820 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.792927 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.792991 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793049 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793094 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793557 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793738 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793790 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793828 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793969 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.793973 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794075 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794109 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794164 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794236 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794641 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70351a6e-b968-4e9c-9d7b-76df8fe7218e-config\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794671 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-etcd-client\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794688 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d6q8\" (UniqueName: \"kubernetes.io/projected/1a4e8be6-660b-4056-92f3-854fadca94d7-kube-api-access-8d6q8\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794708 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/832fdd80-29c7-4bd3-95e7-157c0464703c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794725 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zsvs\" (UniqueName: \"kubernetes.io/projected/70351a6e-b968-4e9c-9d7b-76df8fe7218e-kube-api-access-5zsvs\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794742 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z628z\" (UniqueName: \"kubernetes.io/projected/763d3813-881f-4c0a-8160-5c0a15e17926-kube-api-access-z628z\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794759 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-node-pullsecrets\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794774 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-etcd-client\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794792 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-audit\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794807 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-etcd-serving-ca\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794824 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/549dbc44-f8b5-406d-9431-f2a382f44635-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794838 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-serving-cert\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794854 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832fdd80-29c7-4bd3-95e7-157c0464703c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794869 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794901 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-serving-cert\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794919 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794937 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-serving-cert\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794963 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jd8m\" (UniqueName: \"kubernetes.io/projected/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-kube-api-access-2jd8m\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.794978 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-encryption-config\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795024 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-service-ca\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795048 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-audit-policies\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795083 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-service-ca-bundle\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795100 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-oauth-serving-cert\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795115 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/70351a6e-b968-4e9c-9d7b-76df8fe7218e-machine-approver-tls\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795132 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwdc9\" (UniqueName: \"kubernetes.io/projected/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-kube-api-access-nwdc9\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795149 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt9lf\" (UniqueName: \"kubernetes.io/projected/033280af-18b4-44cb-bc49-91e77308a685-kube-api-access-jt9lf\") pod \"downloads-7954f5f757-wshgn\" (UID: \"033280af-18b4-44cb-bc49-91e77308a685\") " pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795164 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1a4e8be6-660b-4056-92f3-854fadca94d7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795186 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-audit-dir\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795201 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqqrl\" (UniqueName: \"kubernetes.io/projected/832fdd80-29c7-4bd3-95e7-157c0464703c-kube-api-access-vqqrl\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795216 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549dbc44-f8b5-406d-9431-f2a382f44635-config\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795453 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795477 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795502 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795526 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttdtx\" (UniqueName: \"kubernetes.io/projected/6fc9cbad-c139-40ac-82c5-b076343c5288-kube-api-access-ttdtx\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795544 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/763d3813-881f-4c0a-8160-5c0a15e17926-console-serving-cert\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795562 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5wwl\" (UniqueName: \"kubernetes.io/projected/cb60fc35-6c25-466c-8738-cb351b60a2eb-kube-api-access-h5wwl\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795577 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-config\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795639 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795695 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795737 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-encryption-config\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795763 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c4e96a48-4a92-473a-856d-85a034e484ff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k7tgh\" (UID: \"c4e96a48-4a92-473a-856d-85a034e484ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795799 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70351a6e-b968-4e9c-9d7b-76df8fe7218e-auth-proxy-config\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795815 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795938 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796030 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796064 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796178 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796320 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.795816 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcs5p\" (UniqueName: \"kubernetes.io/projected/c4e96a48-4a92-473a-856d-85a034e484ff-kube-api-access-bcs5p\") pod \"cluster-samples-operator-665b6dd947-k7tgh\" (UID: \"c4e96a48-4a92-473a-856d-85a034e484ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796426 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796450 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-serving-cert\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796471 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-image-import-ca\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796495 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-console-config\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796512 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-config\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796527 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-config\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796547 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/763d3813-881f-4c0a-8160-5c0a15e17926-console-oauth-config\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796564 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9cbad-c139-40ac-82c5-b076343c5288-serving-cert\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796586 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw5mz\" (UniqueName: \"kubernetes.io/projected/549dbc44-f8b5-406d-9431-f2a382f44635-kube-api-access-nw5mz\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796619 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g2bg\" (UniqueName: \"kubernetes.io/projected/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-kube-api-access-8g2bg\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796636 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/549dbc44-f8b5-406d-9431-f2a382f44635-images\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796640 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796653 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-trusted-ca\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796672 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cb60fc35-6c25-466c-8738-cb351b60a2eb-audit-dir\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796689 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-trusted-ca-bundle\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796709 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a4e8be6-660b-4056-92f3-854fadca94d7-serving-cert\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796724 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796739 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-config\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796753 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-client-ca\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796760 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796853 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796935 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.797021 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.797090 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.796938 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.797161 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.797235 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.799026 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mhc7m"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.804658 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.808400 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.812053 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.813507 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.813593 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.813916 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.813949 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.813974 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.815640 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.815951 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.816809 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.817257 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.817442 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.818056 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.818726 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.819620 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.833061 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.837575 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.838196 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.839795 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kdkmv"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.840281 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.841554 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.842219 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.842401 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.842965 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.843574 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b5fj6"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.843923 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.849379 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.850133 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.850974 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqw26"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.853536 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.855226 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.857162 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.860952 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.860968 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gsnrn"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.870818 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kbqpg"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.872253 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.874725 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.876619 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nthd7"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.877922 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nctzt"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.878082 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.881628 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-l6mzq"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.882413 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.886652 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.888329 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mmm9"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.889906 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.892705 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.892729 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.893663 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vb7pw"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.895012 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.896956 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kwffm"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.897830 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.898757 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a4e8be6-660b-4056-92f3-854fadca94d7-serving-cert\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.898797 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.898838 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-config\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.898914 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-client-ca\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899116 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5fe047f1-e145-4bb0-b68b-881dce6b51ba-metrics-tls\") pod \"dns-operator-744455d44c-pmg89\" (UID: \"5fe047f1-e145-4bb0-b68b-881dce6b51ba\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899146 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-etcd-client\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899162 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70351a6e-b968-4e9c-9d7b-76df8fe7218e-config\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899180 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/832fdd80-29c7-4bd3-95e7-157c0464703c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899196 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zsvs\" (UniqueName: \"kubernetes.io/projected/70351a6e-b968-4e9c-9d7b-76df8fe7218e-kube-api-access-5zsvs\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899211 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z628z\" (UniqueName: \"kubernetes.io/projected/763d3813-881f-4c0a-8160-5c0a15e17926-kube-api-access-z628z\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899231 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d6q8\" (UniqueName: \"kubernetes.io/projected/1a4e8be6-660b-4056-92f3-854fadca94d7-kube-api-access-8d6q8\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899249 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-node-pullsecrets\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899264 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-etcd-client\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899277 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-audit\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899333 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-etcd-serving-ca\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899354 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/549dbc44-f8b5-406d-9431-f2a382f44635-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899411 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-serving-cert\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899458 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832fdd80-29c7-4bd3-95e7-157c0464703c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899494 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899536 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-serving-cert\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899573 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmbpn\" (UniqueName: \"kubernetes.io/projected/d99c04ca-9c5a-4bac-a833-2f4a51487a04-kube-api-access-bmbpn\") pod \"migrator-59844c95c7-n9672\" (UID: \"d99c04ca-9c5a-4bac-a833-2f4a51487a04\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899575 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899845 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.899896 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtcjb\" (UniqueName: \"kubernetes.io/projected/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-kube-api-access-jtcjb\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900021 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-serving-cert\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900126 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jd8m\" (UniqueName: \"kubernetes.io/projected/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-kube-api-access-2jd8m\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900148 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kt94\" (UniqueName: \"kubernetes.io/projected/008bec79-d4f4-491d-9a93-3058568740e9-kube-api-access-8kt94\") pod \"package-server-manager-789f6589d5-d5bhj\" (UID: \"008bec79-d4f4-491d-9a93-3058568740e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900171 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-encryption-config\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900189 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-service-ca\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900217 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-audit-policies\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900235 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-service-ca-bundle\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900255 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-proxy-tls\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900253 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832fdd80-29c7-4bd3-95e7-157c0464703c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900272 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/70351a6e-b968-4e9c-9d7b-76df8fe7218e-machine-approver-tls\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900294 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-oauth-serving-cert\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900313 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/008bec79-d4f4-491d-9a93-3058568740e9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-d5bhj\" (UID: \"008bec79-d4f4-491d-9a93-3058568740e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900333 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900352 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwdc9\" (UniqueName: \"kubernetes.io/projected/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-kube-api-access-nwdc9\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900370 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-audit-dir\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900391 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqqrl\" (UniqueName: \"kubernetes.io/projected/832fdd80-29c7-4bd3-95e7-157c0464703c-kube-api-access-vqqrl\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900409 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549dbc44-f8b5-406d-9431-f2a382f44635-config\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900426 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt9lf\" (UniqueName: \"kubernetes.io/projected/033280af-18b4-44cb-bc49-91e77308a685-kube-api-access-jt9lf\") pod \"downloads-7954f5f757-wshgn\" (UID: \"033280af-18b4-44cb-bc49-91e77308a685\") " pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900442 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1a4e8be6-660b-4056-92f3-854fadca94d7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900465 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900483 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttdtx\" (UniqueName: \"kubernetes.io/projected/6fc9cbad-c139-40ac-82c5-b076343c5288-kube-api-access-ttdtx\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900501 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/763d3813-881f-4c0a-8160-5c0a15e17926-console-serving-cert\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900519 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5wwl\" (UniqueName: \"kubernetes.io/projected/cb60fc35-6c25-466c-8738-cb351b60a2eb-kube-api-access-h5wwl\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900536 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-images\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900565 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70351a6e-b968-4e9c-9d7b-76df8fe7218e-auth-proxy-config\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900584 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-config\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900599 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-encryption-config\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900607 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-node-pullsecrets\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900617 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c4e96a48-4a92-473a-856d-85a034e484ff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k7tgh\" (UID: \"c4e96a48-4a92-473a-856d-85a034e484ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900681 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcs5p\" (UniqueName: \"kubernetes.io/projected/c4e96a48-4a92-473a-856d-85a034e484ff-kube-api-access-bcs5p\") pod \"cluster-samples-operator-665b6dd947-k7tgh\" (UID: \"c4e96a48-4a92-473a-856d-85a034e484ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900705 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-image-import-ca\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900725 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-console-config\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900743 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-config\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900760 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-serving-cert\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900782 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-config\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900803 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/763d3813-881f-4c0a-8160-5c0a15e17926-console-oauth-config\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900818 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9cbad-c139-40ac-82c5-b076343c5288-serving-cert\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900840 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw5mz\" (UniqueName: \"kubernetes.io/projected/549dbc44-f8b5-406d-9431-f2a382f44635-kube-api-access-nw5mz\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900862 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtl26\" (UniqueName: \"kubernetes.io/projected/5fe047f1-e145-4bb0-b68b-881dce6b51ba-kube-api-access-mtl26\") pod \"dns-operator-744455d44c-pmg89\" (UID: \"5fe047f1-e145-4bb0-b68b-881dce6b51ba\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900914 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g2bg\" (UniqueName: \"kubernetes.io/projected/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-kube-api-access-8g2bg\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900935 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/549dbc44-f8b5-406d-9431-f2a382f44635-images\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900957 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-trusted-ca-bundle\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900971 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-trusted-ca\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900988 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cb60fc35-6c25-466c-8738-cb351b60a2eb-audit-dir\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.901062 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cb60fc35-6c25-466c-8738-cb351b60a2eb-audit-dir\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.901273 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-config\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.901838 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-client-ca\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.901905 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.901904 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.902099 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.902612 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.903795 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.904320 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cb60fc35-6c25-466c-8738-cb351b60a2eb-audit-policies\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.904331 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-etcd-client\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.900364 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-etcd-serving-ca\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.904317 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/832fdd80-29c7-4bd3-95e7-157c0464703c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.904919 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-serving-cert\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.905064 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c4e96a48-4a92-473a-856d-85a034e484ff-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k7tgh\" (UID: \"c4e96a48-4a92-473a-856d-85a034e484ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.905178 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70351a6e-b968-4e9c-9d7b-76df8fe7218e-config\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.905396 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-oauth-serving-cert\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.905426 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-serving-cert\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.905795 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-service-ca\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.905946 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-config\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.906208 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/549dbc44-f8b5-406d-9431-f2a382f44635-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.906291 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/70351a6e-b968-4e9c-9d7b-76df8fe7218e-auth-proxy-config\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908196 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549dbc44-f8b5-406d-9431-f2a382f44635-config\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.906549 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908255 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wshgn"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908269 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pmg89"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.906508 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-etcd-client\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908527 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908544 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cb60fc35-6c25-466c-8738-cb351b60a2eb-encryption-config\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908752 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-audit-dir\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.908974 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/763d3813-881f-4c0a-8160-5c0a15e17926-console-oauth-config\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909171 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/763d3813-881f-4c0a-8160-5c0a15e17926-console-serving-cert\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909308 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1a4e8be6-660b-4056-92f3-854fadca94d7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909458 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/549dbc44-f8b5-406d-9431-f2a382f44635-images\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909497 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.907608 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-trusted-ca\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909530 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9cbad-c139-40ac-82c5-b076343c5288-serving-cert\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909993 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-config\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.909998 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-service-ca-bundle\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.910338 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-console-config\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.910382 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-image-import-ca\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.910496 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-config\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.910510 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a4e8be6-660b-4056-92f3-854fadca94d7-serving-cert\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.910537 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-audit\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.910673 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763d3813-881f-4c0a-8160-5c0a15e17926-trusted-ca-bundle\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.911022 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/70351a6e-b968-4e9c-9d7b-76df8fe7218e-machine-approver-tls\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.911425 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.911429 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-serving-cert\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.911734 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-encryption-config\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.912446 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.912799 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.913425 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-serving-cert\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.913449 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nncgs"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.914475 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.915560 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kdkmv"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.916757 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.917739 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mhc7m"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.918717 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gwt8v"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.919631 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.919969 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-pwnws"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.921131 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.921599 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.923138 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kbqpg"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.930738 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqw26"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.933817 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.935806 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.937045 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nthd7"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.938078 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.939153 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-447c5"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.940190 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.941204 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b5fj6"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.942431 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.943573 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gwt8v"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.945194 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8"] Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.953745 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.972830 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 22 14:29:33 crc kubenswrapper[4914]: I1122 14:29:33.992924 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001431 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5fe047f1-e145-4bb0-b68b-881dce6b51ba-metrics-tls\") pod \"dns-operator-744455d44c-pmg89\" (UID: \"5fe047f1-e145-4bb0-b68b-881dce6b51ba\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001483 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmbpn\" (UniqueName: \"kubernetes.io/projected/d99c04ca-9c5a-4bac-a833-2f4a51487a04-kube-api-access-bmbpn\") pod \"migrator-59844c95c7-n9672\" (UID: \"d99c04ca-9c5a-4bac-a833-2f4a51487a04\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001506 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtcjb\" (UniqueName: \"kubernetes.io/projected/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-kube-api-access-jtcjb\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001542 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kt94\" (UniqueName: \"kubernetes.io/projected/008bec79-d4f4-491d-9a93-3058568740e9-kube-api-access-8kt94\") pod \"package-server-manager-789f6589d5-d5bhj\" (UID: \"008bec79-d4f4-491d-9a93-3058568740e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001568 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-proxy-tls\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001597 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/008bec79-d4f4-491d-9a93-3058568740e9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-d5bhj\" (UID: \"008bec79-d4f4-491d-9a93-3058568740e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001618 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001670 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-images\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.001725 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtl26\" (UniqueName: \"kubernetes.io/projected/5fe047f1-e145-4bb0-b68b-881dce6b51ba-kube-api-access-mtl26\") pod \"dns-operator-744455d44c-pmg89\" (UID: \"5fe047f1-e145-4bb0-b68b-881dce6b51ba\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.002358 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-images\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.002450 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.004561 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-proxy-tls\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.013214 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.036861 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.053132 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.078909 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.092792 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.113213 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.132769 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.154239 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.173621 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.186684 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/008bec79-d4f4-491d-9a93-3058568740e9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-d5bhj\" (UID: \"008bec79-d4f4-491d-9a93-3058568740e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.192715 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.213564 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.233217 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.252829 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.273658 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.293502 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.313709 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.325511 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5fe047f1-e145-4bb0-b68b-881dce6b51ba-metrics-tls\") pod \"dns-operator-744455d44c-pmg89\" (UID: \"5fe047f1-e145-4bb0-b68b-881dce6b51ba\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.332866 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.352723 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.373432 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.393301 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.412912 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.434318 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.453751 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.473872 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.534617 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.553960 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.573220 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.597716 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.613824 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.632677 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.653245 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.673617 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.693611 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.713627 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.733650 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.752569 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.772918 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.793288 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.813612 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.833039 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.851258 4914 request.go:700] Waited for 1.010801663s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.852540 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.873868 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.893728 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.920030 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.933535 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.953280 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.973227 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 14:29:34 crc kubenswrapper[4914]: I1122 14:29:34.992847 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.013184 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.032919 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.053009 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.073842 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.093286 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.114183 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.141850 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.153624 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.173201 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.192573 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.213583 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.232632 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.262248 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.273596 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.293288 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.314457 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.332917 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.353381 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.374483 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.393836 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.414763 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.434032 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.453865 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.473603 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.493955 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.513831 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.533269 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.553600 4914 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.573600 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.621600 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zsvs\" (UniqueName: \"kubernetes.io/projected/70351a6e-b968-4e9c-9d7b-76df8fe7218e-kube-api-access-5zsvs\") pod \"machine-approver-56656f9798-w6692\" (UID: \"70351a6e-b968-4e9c-9d7b-76df8fe7218e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.633453 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z628z\" (UniqueName: \"kubernetes.io/projected/763d3813-881f-4c0a-8160-5c0a15e17926-kube-api-access-z628z\") pod \"console-f9d7485db-gsnrn\" (UID: \"763d3813-881f-4c0a-8160-5c0a15e17926\") " pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.646254 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.660185 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d6q8\" (UniqueName: \"kubernetes.io/projected/1a4e8be6-660b-4056-92f3-854fadca94d7-kube-api-access-8d6q8\") pod \"openshift-config-operator-7777fb866f-kwffm\" (UID: \"1a4e8be6-660b-4056-92f3-854fadca94d7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.663091 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.678618 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqqrl\" (UniqueName: \"kubernetes.io/projected/832fdd80-29c7-4bd3-95e7-157c0464703c-kube-api-access-vqqrl\") pod \"openshift-controller-manager-operator-756b6f6bc6-kqwz5\" (UID: \"832fdd80-29c7-4bd3-95e7-157c0464703c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.701659 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt9lf\" (UniqueName: \"kubernetes.io/projected/033280af-18b4-44cb-bc49-91e77308a685-kube-api-access-jt9lf\") pod \"downloads-7954f5f757-wshgn\" (UID: \"033280af-18b4-44cb-bc49-91e77308a685\") " pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.717914 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" event={"ID":"70351a6e-b968-4e9c-9d7b-76df8fe7218e","Type":"ContainerStarted","Data":"702db7948efc6fd235f08902a06227a7e65e377b77fa17d967b931bb95b11518"} Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.722828 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw5mz\" (UniqueName: \"kubernetes.io/projected/549dbc44-f8b5-406d-9431-f2a382f44635-kube-api-access-nw5mz\") pod \"machine-api-operator-5694c8668f-vb7pw\" (UID: \"549dbc44-f8b5-406d-9431-f2a382f44635\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.732433 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwdc9\" (UniqueName: \"kubernetes.io/projected/b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9-kube-api-access-nwdc9\") pod \"authentication-operator-69f744f599-nctzt\" (UID: \"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.759449 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g2bg\" (UniqueName: \"kubernetes.io/projected/3984e8b2-f1ca-48c7-b661-c0f97d7a513c-kube-api-access-8g2bg\") pod \"apiserver-76f77b778f-vhs7k\" (UID: \"3984e8b2-f1ca-48c7-b661-c0f97d7a513c\") " pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.769444 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.780467 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jd8m\" (UniqueName: \"kubernetes.io/projected/1bda94c0-7ba2-4d14-aa93-38a74ffc0836-kube-api-access-2jd8m\") pod \"console-operator-58897d9998-l6mzq\" (UID: \"1bda94c0-7ba2-4d14-aa93-38a74ffc0836\") " pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.791729 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5wwl\" (UniqueName: \"kubernetes.io/projected/cb60fc35-6c25-466c-8738-cb351b60a2eb-kube-api-access-h5wwl\") pod \"apiserver-7bbb656c7d-dk5kx\" (UID: \"cb60fc35-6c25-466c-8738-cb351b60a2eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.806466 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.815157 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttdtx\" (UniqueName: \"kubernetes.io/projected/6fc9cbad-c139-40ac-82c5-b076343c5288-kube-api-access-ttdtx\") pod \"route-controller-manager-6576b87f9c-bpwxr\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.827098 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.833478 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.837376 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcs5p\" (UniqueName: \"kubernetes.io/projected/c4e96a48-4a92-473a-856d-85a034e484ff-kube-api-access-bcs5p\") pod \"cluster-samples-operator-665b6dd947-k7tgh\" (UID: \"c4e96a48-4a92-473a-856d-85a034e484ff\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.851732 4914 request.go:700] Waited for 1.931634816s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-dockercfg-jwfmh&limit=500&resourceVersion=0 Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.853733 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.872978 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.877379 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.893645 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.912107 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.912301 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kwffm"] Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.912678 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.913246 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.918952 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.929885 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.934384 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.934478 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 22 14:29:35 crc kubenswrapper[4914]: W1122 14:29:35.951662 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a4e8be6_660b_4056_92f3_854fadca94d7.slice/crio-ac61a19a1907ed22eed221766c79629df51df81a84779b0cc22817e2de604e13 WatchSource:0}: Error finding container ac61a19a1907ed22eed221766c79629df51df81a84779b0cc22817e2de604e13: Status 404 returned error can't find the container with id ac61a19a1907ed22eed221766c79629df51df81a84779b0cc22817e2de604e13 Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.952798 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.970767 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmbpn\" (UniqueName: \"kubernetes.io/projected/d99c04ca-9c5a-4bac-a833-2f4a51487a04-kube-api-access-bmbpn\") pod \"migrator-59844c95c7-n9672\" (UID: \"d99c04ca-9c5a-4bac-a833-2f4a51487a04\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" Nov 22 14:29:35 crc kubenswrapper[4914]: I1122 14:29:35.990786 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kt94\" (UniqueName: \"kubernetes.io/projected/008bec79-d4f4-491d-9a93-3058568740e9-kube-api-access-8kt94\") pod \"package-server-manager-789f6589d5-d5bhj\" (UID: \"008bec79-d4f4-491d-9a93-3058568740e9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.005942 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-vhs7k"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.031384 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtcjb\" (UniqueName: \"kubernetes.io/projected/1ccab6a9-aeff-4351-8dd2-f07ecccd5643-kube-api-access-jtcjb\") pod \"machine-config-operator-74547568cd-bjgxs\" (UID: \"1ccab6a9-aeff-4351-8dd2-f07ecccd5643\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.042669 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtl26\" (UniqueName: \"kubernetes.io/projected/5fe047f1-e145-4bb0-b68b-881dce6b51ba-kube-api-access-mtl26\") pod \"dns-operator-744455d44c-pmg89\" (UID: \"5fe047f1-e145-4bb0-b68b-881dce6b51ba\") " pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.048993 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.063593 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gsnrn"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.064013 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.074265 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.088853 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" Nov 22 14:29:36 crc kubenswrapper[4914]: W1122 14:29:36.104436 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod763d3813_881f_4c0a_8160_5c0a15e17926.slice/crio-8d1880fa7dbded51274efda52103d6650a4573241d23e49a4a805b7ec3310f3c WatchSource:0}: Error finding container 8d1880fa7dbded51274efda52103d6650a4573241d23e49a4a805b7ec3310f3c: Status 404 returned error can't find the container with id 8d1880fa7dbded51274efda52103d6650a4573241d23e49a4a805b7ec3310f3c Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.116665 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130599 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130632 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baad2152-3fdb-4a99-baee-17471ae8dfc5-service-ca-bundle\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130654 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130686 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/632b02d3-3730-4ccf-9801-b5e039be071d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130703 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130720 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130745 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98x5t\" (UniqueName: \"kubernetes.io/projected/baad2152-3fdb-4a99-baee-17471ae8dfc5-kube-api-access-98x5t\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130761 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-dir\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130781 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b02d3-3730-4ccf-9801-b5e039be071d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130795 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-metrics-certs\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130812 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-default-certificate\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130846 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130885 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-stats-auth\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130902 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d933ef10-c921-45ea-906d-feaf26e9b4d0-serving-cert\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130941 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-ca\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130958 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.130974 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-service-ca\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131000 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-bound-sa-token\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131018 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131046 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-config\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131062 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-policies\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131080 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lqr7\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-kube-api-access-2lqr7\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131107 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131123 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/efa47bdc-1106-455f-a01a-234f398947ba-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131138 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131156 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrcfv\" (UniqueName: \"kubernetes.io/projected/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-kube-api-access-mrcfv\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131173 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131189 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131234 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-trusted-ca\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131250 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131268 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f152ee7-771c-47ae-bc31-6e4148a017ef-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131301 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131369 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vvfv\" (UniqueName: \"kubernetes.io/projected/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-kube-api-access-7vvfv\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131413 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/eaa079dd-220a-4941-90e7-958571a207bd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nk4jg\" (UID: \"eaa079dd-220a-4941-90e7-958571a207bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131439 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa47bdc-1106-455f-a01a-234f398947ba-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131455 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-client\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131469 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131509 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh2zr\" (UniqueName: \"kubernetes.io/projected/eaa079dd-220a-4941-90e7-958571a207bd-kube-api-access-jh2zr\") pod \"control-plane-machine-set-operator-78cbb6b69f-nk4jg\" (UID: \"eaa079dd-220a-4941-90e7-958571a207bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131555 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f152ee7-771c-47ae-bc31-6e4148a017ef-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131575 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.131637 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.133744 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv88v\" (UniqueName: \"kubernetes.io/projected/632b02d3-3730-4ccf-9801-b5e039be071d-kube-api-access-fv88v\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.133809 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-certificates\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.133831 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-config\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.134144 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efa47bdc-1106-455f-a01a-234f398947ba-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.134164 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjbx9\" (UniqueName: \"kubernetes.io/projected/d933ef10-c921-45ea-906d-feaf26e9b4d0-kube-api-access-hjbx9\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.134264 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-tls\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.134283 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.134859 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:36.634837515 +0000 UTC m=+142.443407176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: W1122 14:29:36.162536 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb60fc35_6c25_466c_8738_cb351b60a2eb.slice/crio-afd71feeb9de504df2fa8fbeadc856fc581f9b83c99944d515448434ccc69c45 WatchSource:0}: Error finding container afd71feeb9de504df2fa8fbeadc856fc581f9b83c99944d515448434ccc69c45: Status 404 returned error can't find the container with id afd71feeb9de504df2fa8fbeadc856fc581f9b83c99944d515448434ccc69c45 Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.234893 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.234964 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:36.734941083 +0000 UTC m=+142.543510744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235289 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baad2152-3fdb-4a99-baee-17471ae8dfc5-service-ca-bundle\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235311 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235337 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/afddd90d-31ac-42e7-9689-dd03a453a983-node-bootstrap-token\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235363 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-config\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235381 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235410 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235428 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235443 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjg4x\" (UniqueName: \"kubernetes.io/projected/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-kube-api-access-vjg4x\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235458 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjs5\" (UniqueName: \"kubernetes.io/projected/afddd90d-31ac-42e7-9689-dd03a453a983-kube-api-access-5gjs5\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235476 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98x5t\" (UniqueName: \"kubernetes.io/projected/baad2152-3fdb-4a99-baee-17471ae8dfc5-kube-api-access-98x5t\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235494 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-dir\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235509 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n86x\" (UniqueName: \"kubernetes.io/projected/9c28db41-412e-415e-b37f-9bf288649379-kube-api-access-2n86x\") pod \"multus-admission-controller-857f4d67dd-mhc7m\" (UID: \"9c28db41-412e-415e-b37f-9bf288649379\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235526 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd24837-1be8-4227-80b3-bd7ea3972277-config\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235543 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmntr\" (UniqueName: \"kubernetes.io/projected/1c100c3f-0c58-462d-9d70-5072422b1822-kube-api-access-cmntr\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235586 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e44ee621-6ddb-4ee6-bf5a-07215d61a280-config-volume\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235603 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9c28db41-412e-415e-b37f-9bf288649379-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mhc7m\" (UID: \"9c28db41-412e-415e-b37f-9bf288649379\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235619 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-plugins-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235638 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-stats-auth\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235679 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-ca\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235696 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-client-ca\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235712 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235736 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-bound-sa-token\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235751 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-service-ca\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235774 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-config\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235790 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/00c2187f-0e88-460b-9f66-92b11fa0f3e7-signing-key\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235808 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5clfs\" (UniqueName: \"kubernetes.io/projected/8fd24837-1be8-4227-80b3-bd7ea3972277-kube-api-access-5clfs\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235826 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lqr7\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-kube-api-access-2lqr7\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235858 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235896 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58d8t\" (UniqueName: \"kubernetes.io/projected/d3e8ec37-2b50-4b6d-8c75-120938956aa7-kube-api-access-58d8t\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235913 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235930 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-serving-cert\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235948 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjwlf\" (UniqueName: \"kubernetes.io/projected/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-kube-api-access-qjwlf\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.235974 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236068 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f152ee7-771c-47ae-bc31-6e4148a017ef-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236127 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vvfv\" (UniqueName: \"kubernetes.io/projected/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-kube-api-access-7vvfv\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236149 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/eaa079dd-220a-4941-90e7-958571a207bd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nk4jg\" (UID: \"eaa079dd-220a-4941-90e7-958571a207bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236186 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-client\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236212 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d3e8ec37-2b50-4b6d-8c75-120938956aa7-metrics-tls\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236246 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd24837-1be8-4227-80b3-bd7ea3972277-serving-cert\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236262 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236337 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236363 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1c100c3f-0c58-462d-9d70-5072422b1822-tmpfs\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236380 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/afddd90d-31ac-42e7-9689-dd03a453a983-certs\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.236467 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baad2152-3fdb-4a99-baee-17471ae8dfc5-service-ca-bundle\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237110 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237775 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36416b65-0ab4-4472-93c8-a1098aee6049-proxy-tls\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237814 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-srv-cert\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237864 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f152ee7-771c-47ae-bc31-6e4148a017ef-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237907 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237930 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-csi-data-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237953 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnf98\" (UniqueName: \"kubernetes.io/projected/073520eb-1db1-49d1-86d0-4a9d2eb703da-kube-api-access-fnf98\") pod \"ingress-canary-kbqpg\" (UID: \"073520eb-1db1-49d1-86d0-4a9d2eb703da\") " pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.237979 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-certificates\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.238041 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-config\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.238079 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv88v\" (UniqueName: \"kubernetes.io/projected/632b02d3-3730-4ccf-9801-b5e039be071d-kube-api-access-fv88v\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.238220 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-dir\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.239630 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f152ee7-771c-47ae-bc31-6e4148a017ef-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.242111 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-service-ca\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.242563 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-config\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.242599 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/eaa079dd-220a-4941-90e7-958571a207bd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nk4jg\" (UID: \"eaa079dd-220a-4941-90e7-958571a207bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.243320 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-ca\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.243983 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-config\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.245403 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248215 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f152ee7-771c-47ae-bc31-6e4148a017ef-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248225 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248642 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-certificates\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248767 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248757 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bts79\" (UniqueName: \"kubernetes.io/projected/434519f8-3fcf-4a39-9335-9c43f6db4163-kube-api-access-bts79\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248856 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.248893 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-mountpoint-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249057 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249082 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/471e388f-ea5a-401d-9116-c81252a59458-trusted-ca\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249148 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/632b02d3-3730-4ccf-9801-b5e039be071d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249169 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249213 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/471e388f-ea5a-401d-9116-c81252a59458-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249246 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdb4n\" (UniqueName: \"kubernetes.io/projected/36416b65-0ab4-4472-93c8-a1098aee6049-kube-api-access-rdb4n\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249300 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmrg8\" (UniqueName: \"kubernetes.io/projected/406cdf66-b685-46a9-b879-89bad870a774-kube-api-access-mmrg8\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249320 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-srv-cert\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249357 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-profile-collector-cert\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249393 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b02d3-3730-4ccf-9801-b5e039be071d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249430 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e44ee621-6ddb-4ee6-bf5a-07215d61a280-secret-volume\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249480 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-default-certificate\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249522 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-metrics-certs\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.249542 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j28bx\" (UniqueName: \"kubernetes.io/projected/00c2187f-0e88-460b-9f66-92b11fa0f3e7-kube-api-access-j28bx\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.251464 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-stats-auth\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.253261 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.253308 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/406cdf66-b685-46a9-b879-89bad870a774-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.253360 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d933ef10-c921-45ea-906d-feaf26e9b4d0-serving-cert\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.253385 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3e8ec37-2b50-4b6d-8c75-120938956aa7-config-volume\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.253445 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d933ef10-c921-45ea-906d-feaf26e9b4d0-etcd-client\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.254002 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b02d3-3730-4ccf-9801-b5e039be071d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.254036 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-socket-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.254389 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.254466 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/471e388f-ea5a-401d-9116-c81252a59458-metrics-tls\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.254556 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-policies\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.254692 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255187 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/efa47bdc-1106-455f-a01a-234f398947ba-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255330 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrcfv\" (UniqueName: \"kubernetes.io/projected/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-kube-api-access-mrcfv\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255361 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c100c3f-0c58-462d-9d70-5072422b1822-apiservice-cert\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255382 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255415 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-trusted-ca\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255436 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255476 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djmrp\" (UniqueName: \"kubernetes.io/projected/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-kube-api-access-djmrp\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255503 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255522 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255540 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa47bdc-1106-455f-a01a-234f398947ba-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255557 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255556 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255578 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/406cdf66-b685-46a9-b879-89bad870a774-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255597 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-config\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255619 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/00c2187f-0e88-460b-9f66-92b11fa0f3e7-signing-cabundle\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255637 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-registration-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255678 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh2zr\" (UniqueName: \"kubernetes.io/projected/eaa079dd-220a-4941-90e7-958571a207bd-kube-api-access-jh2zr\") pod \"control-plane-machine-set-operator-78cbb6b69f-nk4jg\" (UID: \"eaa079dd-220a-4941-90e7-958571a207bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255698 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255717 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqwpc\" (UniqueName: \"kubernetes.io/projected/e44ee621-6ddb-4ee6-bf5a-07215d61a280-kube-api-access-wqwpc\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255743 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c100c3f-0c58-462d-9d70-5072422b1822-webhook-cert\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255787 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efa47bdc-1106-455f-a01a-234f398947ba-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255806 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjbx9\" (UniqueName: \"kubernetes.io/projected/d933ef10-c921-45ea-906d-feaf26e9b4d0-kube-api-access-hjbx9\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255822 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/073520eb-1db1-49d1-86d0-4a9d2eb703da-cert\") pod \"ingress-canary-kbqpg\" (UID: \"073520eb-1db1-49d1-86d0-4a9d2eb703da\") " pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255839 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slnt5\" (UniqueName: \"kubernetes.io/projected/471e388f-ea5a-401d-9116-c81252a59458-kube-api-access-slnt5\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.255902 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-tls\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.256236 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:36.756223563 +0000 UTC m=+142.564793224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.256630 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efa47bdc-1106-455f-a01a-234f398947ba-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.257046 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-trusted-ca\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.261707 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-policies\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.262082 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.262269 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36416b65-0ab4-4472-93c8-a1098aee6049-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.262308 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkck5\" (UniqueName: \"kubernetes.io/projected/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-kube-api-access-zkck5\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.265639 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.267625 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.268712 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efa47bdc-1106-455f-a01a-234f398947ba-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.270988 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-metrics-certs\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.271178 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/632b02d3-3730-4ccf-9801-b5e039be071d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.271189 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.272843 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.273861 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.273987 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/baad2152-3fdb-4a99-baee-17471ae8dfc5-default-certificate\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.274833 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d933ef10-c921-45ea-906d-feaf26e9b4d0-serving-cert\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.275099 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.275242 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.275666 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.276268 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-bound-sa-token\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.278365 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-tls\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.288610 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lqr7\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-kube-api-access-2lqr7\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.316991 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98x5t\" (UniqueName: \"kubernetes.io/projected/baad2152-3fdb-4a99-baee-17471ae8dfc5-kube-api-access-98x5t\") pod \"router-default-5444994796-xfwmr\" (UID: \"baad2152-3fdb-4a99-baee-17471ae8dfc5\") " pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.346503 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.363979 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.364187 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:36.864146507 +0000 UTC m=+142.672716168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364235 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9c28db41-412e-415e-b37f-9bf288649379-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mhc7m\" (UID: \"9c28db41-412e-415e-b37f-9bf288649379\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364273 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-plugins-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364302 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-client-ca\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364330 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5clfs\" (UniqueName: \"kubernetes.io/projected/8fd24837-1be8-4227-80b3-bd7ea3972277-kube-api-access-5clfs\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364386 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/00c2187f-0e88-460b-9f66-92b11fa0f3e7-signing-key\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364416 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58d8t\" (UniqueName: \"kubernetes.io/projected/d3e8ec37-2b50-4b6d-8c75-120938956aa7-kube-api-access-58d8t\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364433 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-serving-cert\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364450 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjwlf\" (UniqueName: \"kubernetes.io/projected/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-kube-api-access-qjwlf\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364474 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364513 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d3e8ec37-2b50-4b6d-8c75-120938956aa7-metrics-tls\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364531 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd24837-1be8-4227-80b3-bd7ea3972277-serving-cert\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364550 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364567 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364594 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1c100c3f-0c58-462d-9d70-5072422b1822-tmpfs\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364614 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/afddd90d-31ac-42e7-9689-dd03a453a983-certs\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364635 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36416b65-0ab4-4472-93c8-a1098aee6049-proxy-tls\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364653 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-srv-cert\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364677 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-csi-data-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364693 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnf98\" (UniqueName: \"kubernetes.io/projected/073520eb-1db1-49d1-86d0-4a9d2eb703da-kube-api-access-fnf98\") pod \"ingress-canary-kbqpg\" (UID: \"073520eb-1db1-49d1-86d0-4a9d2eb703da\") " pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364714 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bts79\" (UniqueName: \"kubernetes.io/projected/434519f8-3fcf-4a39-9335-9c43f6db4163-kube-api-access-bts79\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364732 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-mountpoint-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364748 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/471e388f-ea5a-401d-9116-c81252a59458-trusted-ca\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364775 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/471e388f-ea5a-401d-9116-c81252a59458-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364805 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdb4n\" (UniqueName: \"kubernetes.io/projected/36416b65-0ab4-4472-93c8-a1098aee6049-kube-api-access-rdb4n\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364826 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmrg8\" (UniqueName: \"kubernetes.io/projected/406cdf66-b685-46a9-b879-89bad870a774-kube-api-access-mmrg8\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364844 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-srv-cert\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364862 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-profile-collector-cert\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364896 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e44ee621-6ddb-4ee6-bf5a-07215d61a280-secret-volume\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364914 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j28bx\" (UniqueName: \"kubernetes.io/projected/00c2187f-0e88-460b-9f66-92b11fa0f3e7-kube-api-access-j28bx\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364932 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/406cdf66-b685-46a9-b879-89bad870a774-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364954 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3e8ec37-2b50-4b6d-8c75-120938956aa7-config-volume\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364969 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-socket-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.364984 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/471e388f-ea5a-401d-9116-c81252a59458-metrics-tls\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365027 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c100c3f-0c58-462d-9d70-5072422b1822-apiservice-cert\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365060 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djmrp\" (UniqueName: \"kubernetes.io/projected/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-kube-api-access-djmrp\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365082 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365099 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365118 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/406cdf66-b685-46a9-b879-89bad870a774-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365134 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-config\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365152 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/00c2187f-0e88-460b-9f66-92b11fa0f3e7-signing-cabundle\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365166 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-registration-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365193 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqwpc\" (UniqueName: \"kubernetes.io/projected/e44ee621-6ddb-4ee6-bf5a-07215d61a280-kube-api-access-wqwpc\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365207 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c100c3f-0c58-462d-9d70-5072422b1822-webhook-cert\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365240 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/073520eb-1db1-49d1-86d0-4a9d2eb703da-cert\") pod \"ingress-canary-kbqpg\" (UID: \"073520eb-1db1-49d1-86d0-4a9d2eb703da\") " pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365257 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36416b65-0ab4-4472-93c8-a1098aee6049-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365272 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slnt5\" (UniqueName: \"kubernetes.io/projected/471e388f-ea5a-401d-9116-c81252a59458-kube-api-access-slnt5\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365292 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkck5\" (UniqueName: \"kubernetes.io/projected/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-kube-api-access-zkck5\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365311 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365340 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/afddd90d-31ac-42e7-9689-dd03a453a983-node-bootstrap-token\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365356 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-config\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365375 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365579 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjg4x\" (UniqueName: \"kubernetes.io/projected/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-kube-api-access-vjg4x\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365596 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n86x\" (UniqueName: \"kubernetes.io/projected/9c28db41-412e-415e-b37f-9bf288649379-kube-api-access-2n86x\") pod \"multus-admission-controller-857f4d67dd-mhc7m\" (UID: \"9c28db41-412e-415e-b37f-9bf288649379\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365612 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gjs5\" (UniqueName: \"kubernetes.io/projected/afddd90d-31ac-42e7-9689-dd03a453a983-kube-api-access-5gjs5\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365635 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd24837-1be8-4227-80b3-bd7ea3972277-config\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365653 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmntr\" (UniqueName: \"kubernetes.io/projected/1c100c3f-0c58-462d-9d70-5072422b1822-kube-api-access-cmntr\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.365674 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e44ee621-6ddb-4ee6-bf5a-07215d61a280-config-volume\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.366012 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:36.8659946 +0000 UTC m=+142.674564261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.366372 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-csi-data-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.366435 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e44ee621-6ddb-4ee6-bf5a-07215d61a280-config-volume\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.366568 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-mountpoint-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.367656 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/471e388f-ea5a-401d-9116-c81252a59458-trusted-ca\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.368002 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d3e8ec37-2b50-4b6d-8c75-120938956aa7-config-volume\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.368386 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.368529 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vvfv\" (UniqueName: \"kubernetes.io/projected/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-kube-api-access-7vvfv\") pod \"oauth-openshift-558db77b4-447c5\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.369263 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/406cdf66-b685-46a9-b879-89bad870a774-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.369487 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-registration-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.370098 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd24837-1be8-4227-80b3-bd7ea3972277-config\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.370402 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/00c2187f-0e88-460b-9f66-92b11fa0f3e7-signing-cabundle\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.371984 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/471e388f-ea5a-401d-9116-c81252a59458-metrics-tls\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.372212 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-srv-cert\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.372924 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-plugins-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.372994 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9c28db41-412e-415e-b37f-9bf288649379-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mhc7m\" (UID: \"9c28db41-412e-415e-b37f-9bf288649379\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.373200 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-client-ca\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.374104 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36416b65-0ab4-4472-93c8-a1098aee6049-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.374221 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.374375 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/434519f8-3fcf-4a39-9335-9c43f6db4163-socket-dir\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.375777 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-config\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.376175 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.376730 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1c100c3f-0c58-462d-9d70-5072422b1822-tmpfs\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.377625 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-config\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.379016 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d3e8ec37-2b50-4b6d-8c75-120938956aa7-metrics-tls\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.380296 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c100c3f-0c58-462d-9d70-5072422b1822-webhook-cert\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.380992 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-profile-collector-cert\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.381287 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e44ee621-6ddb-4ee6-bf5a-07215d61a280-secret-volume\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.381326 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd24837-1be8-4227-80b3-bd7ea3972277-serving-cert\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.381930 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/073520eb-1db1-49d1-86d0-4a9d2eb703da-cert\") pod \"ingress-canary-kbqpg\" (UID: \"073520eb-1db1-49d1-86d0-4a9d2eb703da\") " pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.382335 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.382495 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c100c3f-0c58-462d-9d70-5072422b1822-apiservice-cert\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.382519 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-serving-cert\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.382524 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36416b65-0ab4-4472-93c8-a1098aee6049-proxy-tls\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.382645 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.382947 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/406cdf66-b685-46a9-b879-89bad870a774-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.383159 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/00c2187f-0e88-460b-9f66-92b11fa0f3e7-signing-key\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.383620 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/afddd90d-31ac-42e7-9689-dd03a453a983-node-bootstrap-token\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.385280 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/afddd90d-31ac-42e7-9689-dd03a453a983-certs\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.388023 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-srv-cert\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.396340 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.396654 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv88v\" (UniqueName: \"kubernetes.io/projected/632b02d3-3730-4ccf-9801-b5e039be071d-kube-api-access-fv88v\") pod \"openshift-apiserver-operator-796bbdcf4f-vlvbr\" (UID: \"632b02d3-3730-4ccf-9801-b5e039be071d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.407983 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-l6mzq"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.409090 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ebacf887-765c-41b7-b7a6-8f87eb1bcc89-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-z6z2j\" (UID: \"ebacf887-765c-41b7-b7a6-8f87eb1bcc89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.417397 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vb7pw"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.428265 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/efa47bdc-1106-455f-a01a-234f398947ba-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xscgk\" (UID: \"efa47bdc-1106-455f-a01a-234f398947ba\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.449314 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrcfv\" (UniqueName: \"kubernetes.io/projected/b20756f3-b4fc-4a40-983d-b95fe5ce6e7c-kube-api-access-mrcfv\") pod \"cluster-image-registry-operator-dc59b4c8b-jb8z9\" (UID: \"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.459572 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.466215 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.466696 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:36.966680136 +0000 UTC m=+142.775249797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.471490 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh2zr\" (UniqueName: \"kubernetes.io/projected/eaa079dd-220a-4941-90e7-958571a207bd-kube-api-access-jh2zr\") pod \"control-plane-machine-set-operator-78cbb6b69f-nk4jg\" (UID: \"eaa079dd-220a-4941-90e7-958571a207bd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.491611 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjbx9\" (UniqueName: \"kubernetes.io/projected/d933ef10-c921-45ea-906d-feaf26e9b4d0-kube-api-access-hjbx9\") pod \"etcd-operator-b45778765-nncgs\" (UID: \"d933ef10-c921-45ea-906d-feaf26e9b4d0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.507038 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.510328 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.520767 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nctzt"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.527122 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wshgn"] Nov 22 14:29:36 crc kubenswrapper[4914]: W1122 14:29:36.528108 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod832fdd80_29c7_4bd3_95e7_157c0464703c.slice/crio-4749f34d71ad3d7500f8c1f685976b763560a847c387e01c7d0347d0b409809c WatchSource:0}: Error finding container 4749f34d71ad3d7500f8c1f685976b763560a847c387e01c7d0347d0b409809c: Status 404 returned error can't find the container with id 4749f34d71ad3d7500f8c1f685976b763560a847c387e01c7d0347d0b409809c Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.530163 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnf98\" (UniqueName: \"kubernetes.io/projected/073520eb-1db1-49d1-86d0-4a9d2eb703da-kube-api-access-fnf98\") pod \"ingress-canary-kbqpg\" (UID: \"073520eb-1db1-49d1-86d0-4a9d2eb703da\") " pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.536143 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.555333 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bts79\" (UniqueName: \"kubernetes.io/projected/434519f8-3fcf-4a39-9335-9c43f6db4163-kube-api-access-bts79\") pod \"csi-hostpathplugin-nthd7\" (UID: \"434519f8-3fcf-4a39-9335-9c43f6db4163\") " pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.567541 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.567833 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.067820155 +0000 UTC m=+142.876389816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.569108 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.572913 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkck5\" (UniqueName: \"kubernetes.io/projected/e9973ff4-5f40-4dc7-83f3-c9052d8b80ec-kube-api-access-zkck5\") pod \"olm-operator-6b444d44fb-c8r5m\" (UID: \"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.583862 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.588711 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j28bx\" (UniqueName: \"kubernetes.io/projected/00c2187f-0e88-460b-9f66-92b11fa0f3e7-kube-api-access-j28bx\") pod \"service-ca-9c57cc56f-b5fj6\" (UID: \"00c2187f-0e88-460b-9f66-92b11fa0f3e7\") " pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.597105 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.606997 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjg4x\" (UniqueName: \"kubernetes.io/projected/6ae95d22-92c8-4c2c-a446-45a7c18d5fb2-kube-api-access-vjg4x\") pod \"catalog-operator-68c6474976-lxkhk\" (UID: \"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.623842 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pmg89"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.635357 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.637175 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.638103 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.641223 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/471e388f-ea5a-401d-9116-c81252a59458-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: W1122 14:29:36.651236 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe047f1_e145_4bb0_b68b_881dce6b51ba.slice/crio-62bfd64cfef82de15cb4b1d9d0043e23c5fc38216678cd8340f500222f7541ab WatchSource:0}: Error finding container 62bfd64cfef82de15cb4b1d9d0043e23c5fc38216678cd8340f500222f7541ab: Status 404 returned error can't find the container with id 62bfd64cfef82de15cb4b1d9d0043e23c5fc38216678cd8340f500222f7541ab Nov 22 14:29:36 crc kubenswrapper[4914]: W1122 14:29:36.654582 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd99c04ca_9c5a_4bac_a833_2f4a51487a04.slice/crio-9efce79706966650248aaee2bfc897e63230ced790fcaf8da8822c60d1810051 WatchSource:0}: Error finding container 9efce79706966650248aaee2bfc897e63230ced790fcaf8da8822c60d1810051: Status 404 returned error can't find the container with id 9efce79706966650248aaee2bfc897e63230ced790fcaf8da8822c60d1810051 Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.656447 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdb4n\" (UniqueName: \"kubernetes.io/projected/36416b65-0ab4-4472-93c8-a1098aee6049-kube-api-access-rdb4n\") pod \"machine-config-controller-84d6567774-jn8vg\" (UID: \"36416b65-0ab4-4472-93c8-a1098aee6049\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.666210 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.669743 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.670421 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.170394695 +0000 UTC m=+142.978964356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.682152 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.685781 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmrg8\" (UniqueName: \"kubernetes.io/projected/406cdf66-b685-46a9-b879-89bad870a774-kube-api-access-mmrg8\") pod \"marketplace-operator-79b997595-kdkmv\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.713839 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjwlf\" (UniqueName: \"kubernetes.io/projected/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-kube-api-access-qjwlf\") pod \"controller-manager-879f6c89f-tqw26\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.715457 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f6e8f652-916e-49e8-8a0e-57bdd5c8e73e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bz7p9\" (UID: \"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.726657 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slnt5\" (UniqueName: \"kubernetes.io/projected/471e388f-ea5a-401d-9116-c81252a59458-kube-api-access-slnt5\") pod \"ingress-operator-5b745b69d9-s9zrl\" (UID: \"471e388f-ea5a-401d-9116-c81252a59458\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.727959 4914 generic.go:334] "Generic (PLEG): container finished" podID="3984e8b2-f1ca-48c7-b661-c0f97d7a513c" containerID="8c5a84062cc8dc97de634d97d8573d83acd0adbffa1ac63436edd2c0edfa6144" exitCode=0 Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.728010 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" event={"ID":"3984e8b2-f1ca-48c7-b661-c0f97d7a513c","Type":"ContainerDied","Data":"8c5a84062cc8dc97de634d97d8573d83acd0adbffa1ac63436edd2c0edfa6144"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.728038 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" event={"ID":"3984e8b2-f1ca-48c7-b661-c0f97d7a513c","Type":"ContainerStarted","Data":"70a2fcafb27ed08af0abfa12a6e5f79aa9331be7e08f6b03ac17fe9757c435d2"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.730070 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" event={"ID":"008bec79-d4f4-491d-9a93-3058568740e9","Type":"ContainerStarted","Data":"aa39671e45af6f44e66f0d76cb277a26539a769b3a252bf1b1c6496e6d2c4b3a"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.743195 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.743249 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.747161 4914 generic.go:334] "Generic (PLEG): container finished" podID="1a4e8be6-660b-4056-92f3-854fadca94d7" containerID="33ff4042dfb313bf464865236d8911ad13a6128c295e000bbd9653c977cd7eb0" exitCode=0 Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.747222 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" event={"ID":"1a4e8be6-660b-4056-92f3-854fadca94d7","Type":"ContainerDied","Data":"33ff4042dfb313bf464865236d8911ad13a6128c295e000bbd9653c977cd7eb0"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.747248 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" event={"ID":"1a4e8be6-660b-4056-92f3-854fadca94d7","Type":"ContainerStarted","Data":"ac61a19a1907ed22eed221766c79629df51df81a84779b0cc22817e2de604e13"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.752580 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.753069 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" event={"ID":"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9","Type":"ContainerStarted","Data":"e9d3a2e69a261233950ae4e939b9cf3d56a462cf033d8672ea7eab70027164bd"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.757517 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.763799 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n86x\" (UniqueName: \"kubernetes.io/projected/9c28db41-412e-415e-b37f-9bf288649379-kube-api-access-2n86x\") pod \"multus-admission-controller-857f4d67dd-mhc7m\" (UID: \"9c28db41-412e-415e-b37f-9bf288649379\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.764042 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.771000 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" event={"ID":"1ccab6a9-aeff-4351-8dd2-f07ecccd5643","Type":"ContainerStarted","Data":"6dfb4f751e808f9228c4c36cbc9ce35e977c4f274626757c66b8c918caefc2a3"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.771792 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.772347 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.272327867 +0000 UTC m=+143.080897538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.780931 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gjs5\" (UniqueName: \"kubernetes.io/projected/afddd90d-31ac-42e7-9689-dd03a453a983-kube-api-access-5gjs5\") pod \"machine-config-server-pwnws\" (UID: \"afddd90d-31ac-42e7-9689-dd03a453a983\") " pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.781433 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" event={"ID":"5fe047f1-e145-4bb0-b68b-881dce6b51ba","Type":"ContainerStarted","Data":"62bfd64cfef82de15cb4b1d9d0043e23c5fc38216678cd8340f500222f7541ab"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.785256 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wshgn" event={"ID":"033280af-18b4-44cb-bc49-91e77308a685","Type":"ContainerStarted","Data":"0214400c44903d5d0c311f8f5657e23acb0f95733ed33f09e2d75135a75553f9"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.785373 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.792362 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" event={"ID":"1bda94c0-7ba2-4d14-aa93-38a74ffc0836","Type":"ContainerStarted","Data":"8cf01da80d52dda0957f4c070c0b2e94a57778aee38a8a935d45ebbee4655e70"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.792407 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" event={"ID":"1bda94c0-7ba2-4d14-aa93-38a74ffc0836","Type":"ContainerStarted","Data":"4b549a8cbd276321ee79d9643cb11aad6a1b6c808800eaf45eab0c7a8eefb103"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.793649 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.794402 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.806425 4914 patch_prober.go:28] interesting pod/console-operator-58897d9998-l6mzq container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.806503 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" podUID="1bda94c0-7ba2-4d14-aa93-38a74ffc0836" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.806751 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58d8t\" (UniqueName: \"kubernetes.io/projected/d3e8ec37-2b50-4b6d-8c75-120938956aa7-kube-api-access-58d8t\") pod \"dns-default-gwt8v\" (UID: \"d3e8ec37-2b50-4b6d-8c75-120938956aa7\") " pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.815757 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.824648 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kbqpg" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.825115 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmntr\" (UniqueName: \"kubernetes.io/projected/1c100c3f-0c58-462d-9d70-5072422b1822-kube-api-access-cmntr\") pod \"packageserver-d55dfcdfc-25f9c\" (UID: \"1c100c3f-0c58-462d-9d70-5072422b1822\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.832413 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" event={"ID":"d99c04ca-9c5a-4bac-a833-2f4a51487a04","Type":"ContainerStarted","Data":"9efce79706966650248aaee2bfc897e63230ced790fcaf8da8822c60d1810051"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.836560 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5clfs\" (UniqueName: \"kubernetes.io/projected/8fd24837-1be8-4227-80b3-bd7ea3972277-kube-api-access-5clfs\") pod \"service-ca-operator-777779d784-5g2kz\" (UID: \"8fd24837-1be8-4227-80b3-bd7ea3972277\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.841186 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.848736 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.855024 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-pwnws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.857657 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqwpc\" (UniqueName: \"kubernetes.io/projected/e44ee621-6ddb-4ee6-bf5a-07215d61a280-kube-api-access-wqwpc\") pod \"collect-profiles-29397015-9smj8\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.879755 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.880839 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.380805236 +0000 UTC m=+143.189374897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.901296 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djmrp\" (UniqueName: \"kubernetes.io/projected/1fc4384d-2a11-4136-9e3e-4bbd4b127ac5-kube-api-access-djmrp\") pod \"kube-storage-version-migrator-operator-b67b599dd-rt7ws\" (UID: \"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.915815 4914 generic.go:334] "Generic (PLEG): container finished" podID="cb60fc35-6c25-466c-8738-cb351b60a2eb" containerID="dbe56c93afb9cf94b8ea8984a07281aaecd1d87698087e74193c73e16d0c4dd3" exitCode=0 Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930729 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" event={"ID":"6fc9cbad-c139-40ac-82c5-b076343c5288","Type":"ContainerStarted","Data":"9c23132ba2155f034d73a62cbba5a44d371a0a3f9547be555ddd31fe14602a74"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930760 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" event={"ID":"549dbc44-f8b5-406d-9431-f2a382f44635","Type":"ContainerStarted","Data":"bad4d4c93cdfea6a152b24e1d5df72972d281a15c5a91e0af55e2e56f69e77c5"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930780 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9"] Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930795 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" event={"ID":"832fdd80-29c7-4bd3-95e7-157c0464703c","Type":"ContainerStarted","Data":"4749f34d71ad3d7500f8c1f685976b763560a847c387e01c7d0347d0b409809c"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930805 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" event={"ID":"c4e96a48-4a92-473a-856d-85a034e484ff","Type":"ContainerStarted","Data":"e2322d5665e2af8bc52d0f740542761323121aed590cf31e57e8519b1152ef69"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930816 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gsnrn" event={"ID":"763d3813-881f-4c0a-8160-5c0a15e17926","Type":"ContainerStarted","Data":"773dd2b8fc72cb7d3ba67a927d2bf8ec685fbfea07fbb79b0062c09fd075e707"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930827 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gsnrn" event={"ID":"763d3813-881f-4c0a-8160-5c0a15e17926","Type":"ContainerStarted","Data":"8d1880fa7dbded51274efda52103d6650a4573241d23e49a4a805b7ec3310f3c"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930837 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xfwmr" event={"ID":"baad2152-3fdb-4a99-baee-17471ae8dfc5","Type":"ContainerStarted","Data":"e05ada9432c2eb2d78ee7905d880ed49680367ebff7e803ca352176247dea856"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930847 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" event={"ID":"cb60fc35-6c25-466c-8738-cb351b60a2eb","Type":"ContainerDied","Data":"dbe56c93afb9cf94b8ea8984a07281aaecd1d87698087e74193c73e16d0c4dd3"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930863 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" event={"ID":"cb60fc35-6c25-466c-8738-cb351b60a2eb","Type":"ContainerStarted","Data":"afd71feeb9de504df2fa8fbeadc856fc581f9b83c99944d515448434ccc69c45"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930893 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" event={"ID":"70351a6e-b968-4e9c-9d7b-76df8fe7218e","Type":"ContainerStarted","Data":"452aa3cbdf91b98407c3fb564951b17c7bdc8a8300dc307966e3a194dd8c4d29"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.930905 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" event={"ID":"70351a6e-b968-4e9c-9d7b-76df8fe7218e","Type":"ContainerStarted","Data":"c68ba76b0b7fae93de431387eae82cfcdf2ac3c7b2dc0dadbcf79e8f413bc0c2"} Nov 22 14:29:36 crc kubenswrapper[4914]: I1122 14:29:36.981388 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:36 crc kubenswrapper[4914]: E1122 14:29:36.981742 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.481726998 +0000 UTC m=+143.290296659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.005470 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.015583 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.023209 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.046638 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.063428 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.082864 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.085267 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.585243096 +0000 UTC m=+143.393812757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.124643 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.183365 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.186275 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.186663 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.686642643 +0000 UTC m=+143.495212304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.262364 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.267607 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.288198 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.288593 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.788574024 +0000 UTC m=+143.597143685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.297654 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr"] Nov 22 14:29:37 crc kubenswrapper[4914]: W1122 14:29:37.356376 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebacf887_765c_41b7_b7a6_8f87eb1bcc89.slice/crio-46a8405e4ffb42ae3a4c75adaca30626424c2b1a399413bc78c2473012d50590 WatchSource:0}: Error finding container 46a8405e4ffb42ae3a4c75adaca30626424c2b1a399413bc78c2473012d50590: Status 404 returned error can't find the container with id 46a8405e4ffb42ae3a4c75adaca30626424c2b1a399413bc78c2473012d50590 Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.389709 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.390594 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.890574157 +0000 UTC m=+143.699143818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.406073 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.423701 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:37 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:37 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:37 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.423750 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.466598 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-xfwmr" podStartSLOduration=122.466584287 podStartE2EDuration="2m2.466584287s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:37.465536896 +0000 UTC m=+143.274106567" watchObservedRunningTime="2025-11-22 14:29:37.466584287 +0000 UTC m=+143.275153948" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.484672 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kdkmv"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.491062 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:37 crc kubenswrapper[4914]: W1122 14:29:37.491581 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod632b02d3_3730_4ccf_9801_b5e039be071d.slice/crio-8b160d904c852c47bf5733406984a13be35ed5c81098bd3d0e7ccbf75353c3dc WatchSource:0}: Error finding container 8b160d904c852c47bf5733406984a13be35ed5c81098bd3d0e7ccbf75353c3dc: Status 404 returned error can't find the container with id 8b160d904c852c47bf5733406984a13be35ed5c81098bd3d0e7ccbf75353c3dc Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.492415 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:37.992394336 +0000 UTC m=+143.800964007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.501390 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.001374644 +0000 UTC m=+143.809944305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.500573 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.514590 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-gsnrn" podStartSLOduration=122.514573682 podStartE2EDuration="2m2.514573682s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:37.512827092 +0000 UTC m=+143.321396753" watchObservedRunningTime="2025-11-22 14:29:37.514573682 +0000 UTC m=+143.323143343" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.522822 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.602196 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.602498 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.102482432 +0000 UTC m=+143.911052083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.610828 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-447c5"] Nov 22 14:29:37 crc kubenswrapper[4914]: W1122 14:29:37.649000 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9973ff4_5f40_4dc7_83f3_c9052d8b80ec.slice/crio-c34ce01cdb7f8e81cf3c7879038d1fb62e74247ce2750994799595595f23ae0f WatchSource:0}: Error finding container c34ce01cdb7f8e81cf3c7879038d1fb62e74247ce2750994799595595f23ae0f: Status 404 returned error can't find the container with id c34ce01cdb7f8e81cf3c7879038d1fb62e74247ce2750994799595595f23ae0f Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.704725 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:37 crc kubenswrapper[4914]: W1122 14:29:37.709462 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3afd1e74_f3ea_4c65_b46c_51156ed6ae48.slice/crio-fadc0c987b740a4025e54b9c01aa0a2eeac6adf5ea8be506dacd792bace5f5de WatchSource:0}: Error finding container fadc0c987b740a4025e54b9c01aa0a2eeac6adf5ea8be506dacd792bace5f5de: Status 404 returned error can't find the container with id fadc0c987b740a4025e54b9c01aa0a2eeac6adf5ea8be506dacd792bace5f5de Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.726273 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.226247009 +0000 UTC m=+144.034816670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.726616 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqw26"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.757138 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.805421 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.811121 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.31105732 +0000 UTC m=+144.119626981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.829940 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nthd7"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.903937 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gwt8v"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.910263 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:37 crc kubenswrapper[4914]: E1122 14:29:37.910721 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.410699776 +0000 UTC m=+144.219269437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.926739 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.942232 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kbqpg"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.955415 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-w6692" podStartSLOduration=122.955394017 podStartE2EDuration="2m2.955394017s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:37.944040102 +0000 UTC m=+143.752609753" watchObservedRunningTime="2025-11-22 14:29:37.955394017 +0000 UTC m=+143.763963678" Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.985828 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nncgs"] Nov 22 14:29:37 crc kubenswrapper[4914]: I1122 14:29:37.999248 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" event={"ID":"632b02d3-3730-4ccf-9801-b5e039be071d","Type":"ContainerStarted","Data":"8b160d904c852c47bf5733406984a13be35ed5c81098bd3d0e7ccbf75353c3dc"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.002700 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" event={"ID":"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec","Type":"ContainerStarted","Data":"c34ce01cdb7f8e81cf3c7879038d1fb62e74247ce2750994799595595f23ae0f"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.004357 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wshgn" event={"ID":"033280af-18b4-44cb-bc49-91e77308a685","Type":"ContainerStarted","Data":"3e6158ac1a3925a32c5231b7f50c5ef6719a8c84041fc8bc92c8ae156689aa85"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.005599 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b5fj6"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.005621 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.006942 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.008560 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" event={"ID":"406cdf66-b685-46a9-b879-89bad870a774","Type":"ContainerStarted","Data":"0576c01e1d290f31cd3ec73b6eb15ba60075089c6b9cb92268fe428c9184ed60"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.010957 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" event={"ID":"c4e96a48-4a92-473a-856d-85a034e484ff","Type":"ContainerStarted","Data":"4836508d590fc8a3817a7542d7179475fae8a7e7cbe5a8e2997d16e4779ece91"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.015142 4914 patch_prober.go:28] interesting pod/downloads-7954f5f757-wshgn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.015199 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wshgn" podUID="033280af-18b4-44cb-bc49-91e77308a685" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.017040 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.017364 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.517348503 +0000 UTC m=+144.325918154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.020715 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" event={"ID":"6fc9cbad-c139-40ac-82c5-b076343c5288","Type":"ContainerStarted","Data":"44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.021700 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.038264 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" event={"ID":"eaa079dd-220a-4941-90e7-958571a207bd","Type":"ContainerStarted","Data":"f29fddfe23d3edd800e92980044cd01d8969e114aac7c02424b3485e4bbf0e30"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.038308 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" event={"ID":"eaa079dd-220a-4941-90e7-958571a207bd","Type":"ContainerStarted","Data":"e9b276d7973f4569caa67751ca6f6ce36e8dbcac767faf24100bd7d5369b3860"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.059286 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.067007 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" podStartSLOduration=123.066993415 podStartE2EDuration="2m3.066993415s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:38.06506616 +0000 UTC m=+143.873635821" watchObservedRunningTime="2025-11-22 14:29:38.066993415 +0000 UTC m=+143.875563076" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.081431 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" event={"ID":"d99c04ca-9c5a-4bac-a833-2f4a51487a04","Type":"ContainerStarted","Data":"3e54296db701c38a34356296bea456b62f0e90dd232f2dfd35ef29cf0626ca3d"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.085387 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" event={"ID":"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c","Type":"ContainerStarted","Data":"1cd2cde9b1cc3c32d560c0a2514a5759c0e0e57c4f684404530cdabe7a4a6f29"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.085414 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" event={"ID":"b20756f3-b4fc-4a40-983d-b95fe5ce6e7c","Type":"ContainerStarted","Data":"85bb585d4db09fa5033ec61560f282d95ee05c0d2f28efee1ea7445111ebdfaf"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.088498 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" event={"ID":"efa47bdc-1106-455f-a01a-234f398947ba","Type":"ContainerStarted","Data":"53cb1f5306e0e3e25664dcbfd8fc6bc2db869d9310323fea9077f5c51b29de41"} Nov 22 14:29:38 crc kubenswrapper[4914]: W1122 14:29:38.089767 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3e8ec37_2b50_4b6d_8c75_120938956aa7.slice/crio-0419a20900eeb50cdc6d697a5fd7d8b513e7434ee42284068ce287ffec12bd0d WatchSource:0}: Error finding container 0419a20900eeb50cdc6d697a5fd7d8b513e7434ee42284068ce287ffec12bd0d: Status 404 returned error can't find the container with id 0419a20900eeb50cdc6d697a5fd7d8b513e7434ee42284068ce287ffec12bd0d Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.096039 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-pwnws" event={"ID":"afddd90d-31ac-42e7-9689-dd03a453a983","Type":"ContainerStarted","Data":"1cb3280ecc4d15665773e215fc7d841e43bb661e0ce9d53c2c54763b59acdc62"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.096082 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-pwnws" event={"ID":"afddd90d-31ac-42e7-9689-dd03a453a983","Type":"ContainerStarted","Data":"745acdbda0a1835ca973391953f0b48af52c23be0230cbb3a7b9e7a4c970394a"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.101499 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" event={"ID":"3afd1e74-f3ea-4c65-b46c-51156ed6ae48","Type":"ContainerStarted","Data":"fadc0c987b740a4025e54b9c01aa0a2eeac6adf5ea8be506dacd792bace5f5de"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.119535 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.120484 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.620465159 +0000 UTC m=+144.429034820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.127062 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.130160 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" event={"ID":"b24f9dbb-b6d5-4865-99a5-3fb0cb5168c9","Type":"ContainerStarted","Data":"9f062c41c1a264b6b5ca35b0beeb27aa2378ac03857f80b078118f1e43ed2fb5"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.151667 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" event={"ID":"ebacf887-765c-41b7-b7a6-8f87eb1bcc89","Type":"ContainerStarted","Data":"46a8405e4ffb42ae3a4c75adaca30626424c2b1a399413bc78c2473012d50590"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.167271 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" event={"ID":"1ccab6a9-aeff-4351-8dd2-f07ecccd5643","Type":"ContainerStarted","Data":"518b4d807f23ad5a5b3da4fca11edcdef936c359f4ba0666c032be49a629caef"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.177926 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mhc7m"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.178839 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" event={"ID":"5fe047f1-e145-4bb0-b68b-881dce6b51ba","Type":"ContainerStarted","Data":"c4b9eb5e1acac1b7a21b0dc07d01035bff8933ac87009d7164f0040c601a46c5"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.189271 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.201971 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.205437 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" event={"ID":"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2","Type":"ContainerStarted","Data":"e76a9f385c23530f102a74856973fd24525e33052ba5ea0ea08c9aa8f73ce2c9"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.207858 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c"] Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.218826 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" event={"ID":"1a4e8be6-660b-4056-92f3-854fadca94d7","Type":"ContainerStarted","Data":"8693fad37db11d744536abb9c14e319f4e942044d887361416ab1b296fbfec1a"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.219291 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.220825 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.221960 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.721945397 +0000 UTC m=+144.530515058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.227011 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xfwmr" event={"ID":"baad2152-3fdb-4a99-baee-17471ae8dfc5","Type":"ContainerStarted","Data":"59ac72a1eebcdd212380d0338849dfcf9f1ccd264829042a76d9abb4d56c96ae"} Nov 22 14:29:38 crc kubenswrapper[4914]: W1122 14:29:38.230351 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fd24837_1be8_4227_80b3_bd7ea3972277.slice/crio-6b2d78d4e9fead5c6e8ab139669ee86983be889e144ffab7cce1c446715b8dc4 WatchSource:0}: Error finding container 6b2d78d4e9fead5c6e8ab139669ee86983be889e144ffab7cce1c446715b8dc4: Status 404 returned error can't find the container with id 6b2d78d4e9fead5c6e8ab139669ee86983be889e144ffab7cce1c446715b8dc4 Nov 22 14:29:38 crc kubenswrapper[4914]: W1122 14:29:38.241514 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c28db41_412e_415e_b37f_9bf288649379.slice/crio-b47a9450377e4361bd39624cb710b2c5a924fd78fbf8d27773b352afe3508383 WatchSource:0}: Error finding container b47a9450377e4361bd39624cb710b2c5a924fd78fbf8d27773b352afe3508383: Status 404 returned error can't find the container with id b47a9450377e4361bd39624cb710b2c5a924fd78fbf8d27773b352afe3508383 Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.241657 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" event={"ID":"4d1efa71-4de9-43b1-8a6c-bcbf686b8057","Type":"ContainerStarted","Data":"c57147b2d91d5186016e67f3e52a943d821206c6c34e5490107461cccdf60189"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.257607 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" event={"ID":"832fdd80-29c7-4bd3-95e7-157c0464703c","Type":"ContainerStarted","Data":"e6ba7f831e3bd60467527061fa12a0815a72424539398b47d5fb5abb0441e693"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.272213 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" event={"ID":"549dbc44-f8b5-406d-9431-f2a382f44635","Type":"ContainerStarted","Data":"0011225459b4663dcde7208db8c930014b377de93c819c5a1007240d44ae6caa"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.272273 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" event={"ID":"549dbc44-f8b5-406d-9431-f2a382f44635","Type":"ContainerStarted","Data":"9bfc87bf9957b5902b7378a900fe740e19a97a13778516ea2ba89895a18d4065"} Nov 22 14:29:38 crc kubenswrapper[4914]: W1122 14:29:38.276164 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode44ee621_6ddb_4ee6_bf5a_07215d61a280.slice/crio-b385cfbac31c51696b659fd6a4fca4e5c193546e609d2ad3ce1b3187c4677694 WatchSource:0}: Error finding container b385cfbac31c51696b659fd6a4fca4e5c193546e609d2ad3ce1b3187c4677694: Status 404 returned error can't find the container with id b385cfbac31c51696b659fd6a4fca4e5c193546e609d2ad3ce1b3187c4677694 Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.279591 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" event={"ID":"3984e8b2-f1ca-48c7-b661-c0f97d7a513c","Type":"ContainerStarted","Data":"fdb06e60f45a8dc870ae36179570b259a4c57d0334e58b0cc7e6dd64b518ec10"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.309805 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" event={"ID":"008bec79-d4f4-491d-9a93-3058568740e9","Type":"ContainerStarted","Data":"7beeffe77fc7d82062641f6c429446341a1fc84324b4eced0b70e3b6d51acfac"} Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.312092 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.324595 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.326150 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-l6mzq" Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.327482 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.827464972 +0000 UTC m=+144.636034633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.409246 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:38 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:38 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:38 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.409294 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.426715 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.428503 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:38.928479807 +0000 UTC m=+144.737049468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.535603 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.536011 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.035997788 +0000 UTC m=+144.844567449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.614638 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-wshgn" podStartSLOduration=123.614621893 podStartE2EDuration="2m3.614621893s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:38.58416255 +0000 UTC m=+144.392732211" watchObservedRunningTime="2025-11-22 14:29:38.614621893 +0000 UTC m=+144.423191554" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.638017 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.638372 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.138354613 +0000 UTC m=+144.946924274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.662949 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vb7pw" podStartSLOduration=123.662934927 podStartE2EDuration="2m3.662934927s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:38.615408275 +0000 UTC m=+144.423977956" watchObservedRunningTime="2025-11-22 14:29:38.662934927 +0000 UTC m=+144.471504588" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.698085 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" podStartSLOduration=123.698069344 podStartE2EDuration="2m3.698069344s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:38.66444579 +0000 UTC m=+144.473015451" watchObservedRunningTime="2025-11-22 14:29:38.698069344 +0000 UTC m=+144.506639005" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.739966 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.740256 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.240244323 +0000 UTC m=+145.048813984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.841007 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.841652 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.34163535 +0000 UTC m=+145.150205011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.896218 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" podStartSLOduration=123.896200044 podStartE2EDuration="2m3.896200044s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:38.892613821 +0000 UTC m=+144.701183482" watchObservedRunningTime="2025-11-22 14:29:38.896200044 +0000 UTC m=+144.704769705" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.902734 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.935091 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" podStartSLOduration=123.935071847 podStartE2EDuration="2m3.935071847s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:38.933235495 +0000 UTC m=+144.741805156" watchObservedRunningTime="2025-11-22 14:29:38.935071847 +0000 UTC m=+144.743641508" Nov 22 14:29:38 crc kubenswrapper[4914]: I1122 14:29:38.942652 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:38 crc kubenswrapper[4914]: E1122 14:29:38.943070 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.443055586 +0000 UTC m=+145.251625247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.043770 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.043971 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.543944358 +0000 UTC m=+145.352514019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.048924 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.049474 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.549454766 +0000 UTC m=+145.358024427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.073535 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kqwz5" podStartSLOduration=124.073517286 podStartE2EDuration="2m4.073517286s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.036089054 +0000 UTC m=+144.844658715" watchObservedRunningTime="2025-11-22 14:29:39.073517286 +0000 UTC m=+144.882086947" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.137474 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" podStartSLOduration=124.137457089 podStartE2EDuration="2m4.137457089s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.106703437 +0000 UTC m=+144.915273098" watchObservedRunningTime="2025-11-22 14:29:39.137457089 +0000 UTC m=+144.946026740" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.151407 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.151849 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.65183167 +0000 UTC m=+145.460401331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.174924 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nctzt" podStartSLOduration=124.174896022 podStartE2EDuration="2m4.174896022s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.171837854 +0000 UTC m=+144.980407515" watchObservedRunningTime="2025-11-22 14:29:39.174896022 +0000 UTC m=+144.983465683" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.239348 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nk4jg" podStartSLOduration=124.239321099 podStartE2EDuration="2m4.239321099s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.238630868 +0000 UTC m=+145.047200529" watchObservedRunningTime="2025-11-22 14:29:39.239321099 +0000 UTC m=+145.047890760" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.253534 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.254336 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.754323708 +0000 UTC m=+145.562893369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.301510 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-pwnws" podStartSLOduration=6.30149229 podStartE2EDuration="6.30149229s" podCreationTimestamp="2025-11-22 14:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.261738091 +0000 UTC m=+145.070307752" watchObservedRunningTime="2025-11-22 14:29:39.30149229 +0000 UTC m=+145.110061951" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.354760 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.354939 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.854908351 +0000 UTC m=+145.663478022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.355681 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.356122 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.856095495 +0000 UTC m=+145.664665156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.392336 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" event={"ID":"d99c04ca-9c5a-4bac-a833-2f4a51487a04","Type":"ContainerStarted","Data":"f3fedf8bc63e3c94b3c12407b219da1f82556bbad560fda4e2f3d959a236667e"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.406311 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" event={"ID":"c4e96a48-4a92-473a-856d-85a034e484ff","Type":"ContainerStarted","Data":"66000ae5ed4e880f8fa9453543ff3c0ea9241dd8a7430e007052f9975d64eb69"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.414687 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:39 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:39 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:39 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.414793 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.415613 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jb8z9" podStartSLOduration=124.415596751 podStartE2EDuration="2m4.415596751s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.414815719 +0000 UTC m=+145.223385390" watchObservedRunningTime="2025-11-22 14:29:39.415596751 +0000 UTC m=+145.224166402" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.434208 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" event={"ID":"e9973ff4-5f40-4dc7-83f3-c9052d8b80ec","Type":"ContainerStarted","Data":"38ef14b24dba5dd387b7dd93bdf8fc2bc83b9e37cfc9b614c551b69a297bb103"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.436730 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.438556 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gwt8v" event={"ID":"d3e8ec37-2b50-4b6d-8c75-120938956aa7","Type":"ContainerStarted","Data":"1e3d5c145883caf22db5c4b6474156b4e9711109c962f0fba125ee9934d65e3b"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.438604 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gwt8v" event={"ID":"d3e8ec37-2b50-4b6d-8c75-120938956aa7","Type":"ContainerStarted","Data":"0419a20900eeb50cdc6d697a5fd7d8b513e7434ee42284068ce287ffec12bd0d"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.441239 4914 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c8r5m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.441286 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" podUID="e9973ff4-5f40-4dc7-83f3-c9052d8b80ec" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.453774 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" event={"ID":"e44ee621-6ddb-4ee6-bf5a-07215d61a280","Type":"ContainerStarted","Data":"b385cfbac31c51696b659fd6a4fca4e5c193546e609d2ad3ce1b3187c4677694"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.458762 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.458975 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.958947794 +0000 UTC m=+145.767517455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.459096 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.459194 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k7tgh" podStartSLOduration=124.45917906 podStartE2EDuration="2m4.45917906s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.458721127 +0000 UTC m=+145.267290788" watchObservedRunningTime="2025-11-22 14:29:39.45917906 +0000 UTC m=+145.267748721" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.460550 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:39.960534989 +0000 UTC m=+145.769104650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.470648 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" event={"ID":"434519f8-3fcf-4a39-9335-9c43f6db4163","Type":"ContainerStarted","Data":"3220cf4ff6d79ce02060ab59bfaaf0af1829ab79d8377c9dd2d428930712ef62"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.489387 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n9672" podStartSLOduration=124.489366845 podStartE2EDuration="2m4.489366845s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.487387489 +0000 UTC m=+145.295957140" watchObservedRunningTime="2025-11-22 14:29:39.489366845 +0000 UTC m=+145.297936506" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.503913 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bjgxs" event={"ID":"1ccab6a9-aeff-4351-8dd2-f07ecccd5643","Type":"ContainerStarted","Data":"feb411509239180a6bd1f60fff343cb5df3bedcb50a979b658303ed5b77b7f7e"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.513948 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" event={"ID":"1c100c3f-0c58-462d-9d70-5072422b1822","Type":"ContainerStarted","Data":"e5dd813845c2ae1b5ef9731de1080e4ab412565ba5112b47587c1623ef48c24c"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.514285 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.515450 4914 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-25f9c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.515494 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" podUID="1c100c3f-0c58-462d-9d70-5072422b1822" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.522046 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kbqpg" event={"ID":"073520eb-1db1-49d1-86d0-4a9d2eb703da","Type":"ContainerStarted","Data":"0c444f2d0ca59d443bc90c83a51c32cc4d85c04e7fe8ced2b7083858b415b229"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.522090 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kbqpg" event={"ID":"073520eb-1db1-49d1-86d0-4a9d2eb703da","Type":"ContainerStarted","Data":"d202c6c996bba30f53a5da0507ed4c7acd6b3b63be9aef6a4a84e0e9cb9e92d4"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.531522 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" event={"ID":"5fe047f1-e145-4bb0-b68b-881dce6b51ba","Type":"ContainerStarted","Data":"ddc38e98f6fca65f8511b44b568c6e7d5578618a069606c5a75a9381d47cd8ca"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.561447 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.562102 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" event={"ID":"cb60fc35-6c25-466c-8738-cb351b60a2eb","Type":"ContainerStarted","Data":"19f9f4a1fb35d087d4454eeb101393b742a33146e7a826d617d839418dca3289"} Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.562781 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.062755479 +0000 UTC m=+145.871325140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.586949 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" event={"ID":"efa47bdc-1106-455f-a01a-234f398947ba","Type":"ContainerStarted","Data":"20365b281bfd6431ff038f57b612627c01c0e0e2beec39912e2b04edacc388e1"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.602746 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" podStartSLOduration=124.602727194 podStartE2EDuration="2m4.602727194s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.601706916 +0000 UTC m=+145.410276577" watchObservedRunningTime="2025-11-22 14:29:39.602727194 +0000 UTC m=+145.411296855" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.603559 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" event={"ID":"632b02d3-3730-4ccf-9801-b5e039be071d","Type":"ContainerStarted","Data":"e59b57123b94fa8019ee0767188a837cb14682302d1ac9eb12be22ff4df53433"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.628112 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kbqpg" podStartSLOduration=6.6280958309999995 podStartE2EDuration="6.628095831s" podCreationTimestamp="2025-11-22 14:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.625633672 +0000 UTC m=+145.434203333" watchObservedRunningTime="2025-11-22 14:29:39.628095831 +0000 UTC m=+145.436665492" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.634939 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" event={"ID":"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e","Type":"ContainerStarted","Data":"a8c14d48f5e0051112a01be54a16e4e71c19467f1864ebedad3a7f05d3c35215"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.654728 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" event={"ID":"6ae95d22-92c8-4c2c-a446-45a7c18d5fb2","Type":"ContainerStarted","Data":"c9123c9996ee6e7c5f275726d1a1d2459a5df19ff6223b3b1ca88cc731a0476e"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.656729 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.660100 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" podStartSLOduration=124.660074778 podStartE2EDuration="2m4.660074778s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.654322103 +0000 UTC m=+145.462891764" watchObservedRunningTime="2025-11-22 14:29:39.660074778 +0000 UTC m=+145.468644439" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.662791 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.664240 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.164223607 +0000 UTC m=+145.972793258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.673317 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.711410 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" podStartSLOduration=124.711377599 podStartE2EDuration="2m4.711377599s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.710308218 +0000 UTC m=+145.518877879" watchObservedRunningTime="2025-11-22 14:29:39.711377599 +0000 UTC m=+145.519947260" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.719476 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" event={"ID":"3984e8b2-f1ca-48c7-b661-c0f97d7a513c","Type":"ContainerStarted","Data":"563ac034815b5a3dddc15c5dd5c1d1dba8177cd8e0431df05396d3b671e5e7cb"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.753616 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" event={"ID":"406cdf66-b685-46a9-b879-89bad870a774","Type":"ContainerStarted","Data":"ee809704a7a393b86e908be334d87adce0e5a56685efccb2b18d5e45990c8558"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.754301 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.755982 4914 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kdkmv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.756052 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.764715 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.764932 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.264901153 +0000 UTC m=+146.073470824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.766271 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" event={"ID":"36416b65-0ab4-4472-93c8-a1098aee6049","Type":"ContainerStarted","Data":"d21b0cec49c3005915ff4ea174fdff9844367fe46f7686edef4de4b5e8ceeee4"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.766331 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" event={"ID":"36416b65-0ab4-4472-93c8-a1098aee6049","Type":"ContainerStarted","Data":"f34ba8d6eedd454e936f78db7b1895f29a8c9320b02ff0bb106c4b145335e4ae"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.767196 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.769155 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.269136344 +0000 UTC m=+146.077706005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.777217 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pmg89" podStartSLOduration=124.777199616 podStartE2EDuration="2m4.777199616s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.77631507 +0000 UTC m=+145.584884731" watchObservedRunningTime="2025-11-22 14:29:39.777199616 +0000 UTC m=+145.585769277" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.777341 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" podStartSLOduration=124.77733574 podStartE2EDuration="2m4.77733574s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.750599293 +0000 UTC m=+145.559168954" watchObservedRunningTime="2025-11-22 14:29:39.77733574 +0000 UTC m=+145.585905401" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.798503 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" event={"ID":"ebacf887-765c-41b7-b7a6-8f87eb1bcc89","Type":"ContainerStarted","Data":"0e65479edfc3e2f1454289ae1a68526d0a623ad38210444fd8323023d3927208"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.802987 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xscgk" podStartSLOduration=124.802962824 podStartE2EDuration="2m4.802962824s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.802231983 +0000 UTC m=+145.610801644" watchObservedRunningTime="2025-11-22 14:29:39.802962824 +0000 UTC m=+145.611532485" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.816044 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" event={"ID":"3afd1e74-f3ea-4c65-b46c-51156ed6ae48","Type":"ContainerStarted","Data":"5ddced9750da27bf2c5e12dd8a983fd2097fa5c2d1bbce94fe2eee14d2e3e1a0"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.816084 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.822550 4914 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-447c5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.30:6443/healthz\": dial tcp 10.217.0.30:6443: connect: connection refused" start-of-body= Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.822599 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" podUID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.30:6443/healthz\": dial tcp 10.217.0.30:6443: connect: connection refused" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.834655 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" event={"ID":"d933ef10-c921-45ea-906d-feaf26e9b4d0","Type":"ContainerStarted","Data":"0b0aec40d80e6246dd16827ac3f1fe55fbd7fa4313264a4c5772f923f6d523d6"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.834698 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" event={"ID":"d933ef10-c921-45ea-906d-feaf26e9b4d0","Type":"ContainerStarted","Data":"b7e9b100dfff8bb0828688674f6ed939d2f2cdf7e8fef0a6257026149b70d81c"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.862549 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vlvbr" podStartSLOduration=124.862533332 podStartE2EDuration="2m4.862533332s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.859528176 +0000 UTC m=+145.668097837" watchObservedRunningTime="2025-11-22 14:29:39.862533332 +0000 UTC m=+145.671102993" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.869864 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.870552 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" event={"ID":"471e388f-ea5a-401d-9116-c81252a59458","Type":"ContainerStarted","Data":"ad3456c405df488673f1f3bac559aa0125c6bba8e70c07f651122fb454029fe4"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.870603 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" event={"ID":"471e388f-ea5a-401d-9116-c81252a59458","Type":"ContainerStarted","Data":"1310fb163ff5401b8464e5dd0b37869dbf987c4f749587f3288310c3802d04bd"} Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.871003 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.370965113 +0000 UTC m=+146.179534774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.872008 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.875390 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.37537383 +0000 UTC m=+146.183943491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.880059 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" event={"ID":"4d1efa71-4de9-43b1-8a6c-bcbf686b8057","Type":"ContainerStarted","Data":"d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.880435 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.881750 4914 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tqw26 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.881791 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" podUID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.901556 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" podStartSLOduration=124.90153047 podStartE2EDuration="2m4.90153047s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.894734885 +0000 UTC m=+145.703304546" watchObservedRunningTime="2025-11-22 14:29:39.90153047 +0000 UTC m=+145.710100131" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.928422 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" event={"ID":"008bec79-d4f4-491d-9a93-3058568740e9","Type":"ContainerStarted","Data":"7c9aaf8e225515958d7980ad05e2c86581d6f73a36ce001c47bc6d6ebdebde26"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.955216 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" event={"ID":"8fd24837-1be8-4227-80b3-bd7ea3972277","Type":"ContainerStarted","Data":"59d2ba17ca1daa87fcf929308fc70baf46764a997cc01624428f5abf945324a6"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.955267 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" event={"ID":"8fd24837-1be8-4227-80b3-bd7ea3972277","Type":"ContainerStarted","Data":"6b2d78d4e9fead5c6e8ab139669ee86983be889e144ffab7cce1c446715b8dc4"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.965406 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lxkhk" podStartSLOduration=124.965378269 podStartE2EDuration="2m4.965378269s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.94867972 +0000 UTC m=+145.757249391" watchObservedRunningTime="2025-11-22 14:29:39.965378269 +0000 UTC m=+145.773947930" Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.973302 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:39 crc kubenswrapper[4914]: E1122 14:29:39.974683 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.474658956 +0000 UTC m=+146.283228617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.991757 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" event={"ID":"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5","Type":"ContainerStarted","Data":"e313e44cd48b0a682f5cc00338b185dd4a4e35e9ad9068e7d4e2fcdb6cf3dd20"} Nov 22 14:29:39 crc kubenswrapper[4914]: I1122 14:29:39.999341 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" event={"ID":"9c28db41-412e-415e-b37f-9bf288649379","Type":"ContainerStarted","Data":"b47a9450377e4361bd39624cb710b2c5a924fd78fbf8d27773b352afe3508383"} Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.004238 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" event={"ID":"00c2187f-0e88-460b-9f66-92b11fa0f3e7","Type":"ContainerStarted","Data":"3077fee66acc83dd5331cfbebc75182c9066c9ac4fa03a4298e0ef5a492c0e3c"} Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.004278 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" event={"ID":"00c2187f-0e88-460b-9f66-92b11fa0f3e7","Type":"ContainerStarted","Data":"37c190b5e949739fbb731519ee24502aed238eccc24e77c103c9410c3b856bb7"} Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.010297 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" podStartSLOduration=125.010250445 podStartE2EDuration="2m5.010250445s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:39.986835975 +0000 UTC m=+145.795405636" watchObservedRunningTime="2025-11-22 14:29:40.010250445 +0000 UTC m=+145.818820106" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.012731 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" podStartSLOduration=125.012724256 podStartE2EDuration="2m5.012724256s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.009184375 +0000 UTC m=+145.817754056" watchObservedRunningTime="2025-11-22 14:29:40.012724256 +0000 UTC m=+145.821293917" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.015498 4914 patch_prober.go:28] interesting pod/downloads-7954f5f757-wshgn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.015528 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wshgn" podUID="033280af-18b4-44cb-bc49-91e77308a685" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.025179 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kwffm" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.046166 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" podStartSLOduration=125.046151905 podStartE2EDuration="2m5.046151905s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.043648653 +0000 UTC m=+145.852218324" watchObservedRunningTime="2025-11-22 14:29:40.046151905 +0000 UTC m=+145.854721566" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.075170 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.076778 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.576765342 +0000 UTC m=+146.385335003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.107586 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" podStartSLOduration=125.107570165 podStartE2EDuration="2m5.107570165s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.101927974 +0000 UTC m=+145.910497625" watchObservedRunningTime="2025-11-22 14:29:40.107570165 +0000 UTC m=+145.916139826" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.128480 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" podStartSLOduration=125.128461994 podStartE2EDuration="2m5.128461994s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.127505076 +0000 UTC m=+145.936074757" watchObservedRunningTime="2025-11-22 14:29:40.128461994 +0000 UTC m=+145.937031665" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.176675 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.177449 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.677432428 +0000 UTC m=+146.486002089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.224451 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-z6z2j" podStartSLOduration=125.224432864 podStartE2EDuration="2m5.224432864s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.224217308 +0000 UTC m=+146.032786979" watchObservedRunningTime="2025-11-22 14:29:40.224432864 +0000 UTC m=+146.033002525" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.225858 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-nncgs" podStartSLOduration=125.225853105 podStartE2EDuration="2m5.225853105s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.176045758 +0000 UTC m=+145.984615419" watchObservedRunningTime="2025-11-22 14:29:40.225853105 +0000 UTC m=+146.034422766" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.260384 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" podStartSLOduration=125.260368615 podStartE2EDuration="2m5.260368615s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.258385388 +0000 UTC m=+146.066955049" watchObservedRunningTime="2025-11-22 14:29:40.260368615 +0000 UTC m=+146.068938276" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.292285 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.292743 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.792729152 +0000 UTC m=+146.601298813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.365080 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-b5fj6" podStartSLOduration=125.365062156 podStartE2EDuration="2m5.365062156s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.361795572 +0000 UTC m=+146.170365233" watchObservedRunningTime="2025-11-22 14:29:40.365062156 +0000 UTC m=+146.173631817" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.394447 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.394587 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.894561221 +0000 UTC m=+146.703130882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.394676 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.395074 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.895057105 +0000 UTC m=+146.703626766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.402223 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:40 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:40 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:40 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.402272 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.404344 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" podStartSLOduration=125.404334661 podStartE2EDuration="2m5.404334661s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.404076114 +0000 UTC m=+146.212645775" watchObservedRunningTime="2025-11-22 14:29:40.404334661 +0000 UTC m=+146.212904322" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.430005 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5g2kz" podStartSLOduration=125.429991577 podStartE2EDuration="2m5.429991577s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:40.428093293 +0000 UTC m=+146.236662954" watchObservedRunningTime="2025-11-22 14:29:40.429991577 +0000 UTC m=+146.238561238" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.495412 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.495655 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:40.995638568 +0000 UTC m=+146.804208229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.596924 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.597199 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.09718882 +0000 UTC m=+146.905758481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.697317 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.697540 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.197510884 +0000 UTC m=+147.006080555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.697637 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.697959 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.197951108 +0000 UTC m=+147.006520769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.770235 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.770283 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.798030 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.798250 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.298214821 +0000 UTC m=+147.106784482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.827956 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.828005 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.836396 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.899611 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:40 crc kubenswrapper[4914]: E1122 14:29:40.899960 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.399945477 +0000 UTC m=+147.208515138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:40 crc kubenswrapper[4914]: I1122 14:29:40.999911 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.000072 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.500054056 +0000 UTC m=+147.308623727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.009242 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gwt8v" event={"ID":"d3e8ec37-2b50-4b6d-8c75-120938956aa7","Type":"ContainerStarted","Data":"0790ae56fcd2dd5ca46dcb3cff9961f3fef8dc1feaf609de0e0e3e81895c05b0"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.009312 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.011602 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s9zrl" event={"ID":"471e388f-ea5a-401d-9116-c81252a59458","Type":"ContainerStarted","Data":"87d3b02dadf2162236726d6cb00b6bc48fd2eb970fe31dd8321d19b450c2721e"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.013063 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" event={"ID":"9c28db41-412e-415e-b37f-9bf288649379","Type":"ContainerStarted","Data":"014637f9b17e26af0ee6f08615076d1e211ec46b8643c45d4c6716bbb64eba25"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.013090 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" event={"ID":"9c28db41-412e-415e-b37f-9bf288649379","Type":"ContainerStarted","Data":"b01d945c41129e50d350385391e5542fe7cbba0ef08530e93b56a1b314d810a9"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.014377 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rt7ws" event={"ID":"1fc4384d-2a11-4136-9e3e-4bbd4b127ac5","Type":"ContainerStarted","Data":"a179a5974f4628a00c98f51bba1df4ebe9ac40fe402a78e7fe8240c4c731c2e5"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.017938 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bz7p9" event={"ID":"f6e8f652-916e-49e8-8a0e-57bdd5c8e73e","Type":"ContainerStarted","Data":"3404c480ad12441f4dc6b027a1be2944dc9c170595abb79fd883f6e88f399fb3"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.052612 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jn8vg" event={"ID":"36416b65-0ab4-4472-93c8-a1098aee6049","Type":"ContainerStarted","Data":"55fb8dc0529d5b9fd3d640a5cf208eb8679083571462705464a778e08f5a2271"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.059682 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" event={"ID":"1c100c3f-0c58-462d-9d70-5072422b1822","Type":"ContainerStarted","Data":"454b097827f08c14cb742bfac9eebb1ba2a9c56c400069d9c0f4a5730330c879"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.061016 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gwt8v" podStartSLOduration=8.061000354 podStartE2EDuration="8.061000354s" podCreationTimestamp="2025-11-22 14:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:41.058285226 +0000 UTC m=+146.866854887" watchObservedRunningTime="2025-11-22 14:29:41.061000354 +0000 UTC m=+146.869570015" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.073341 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" event={"ID":"e44ee621-6ddb-4ee6-bf5a-07215d61a280","Type":"ContainerStarted","Data":"e5bfe0155e6c48f5dae83c44bf076f529cfa1827f58e5b7f2300e3a0de019c2b"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.081769 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" event={"ID":"434519f8-3fcf-4a39-9335-9c43f6db4163","Type":"ContainerStarted","Data":"a7880c25be9bbd0a2078ba3d791c7f582fe5e4ae45ea946c16cc38b43efd7b6a"} Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.082802 4914 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kdkmv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.082900 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.084068 4914 patch_prober.go:28] interesting pod/downloads-7954f5f757-wshgn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.084097 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wshgn" podUID="033280af-18b4-44cb-bc49-91e77308a685" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.095361 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.101061 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.101647 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dk5kx" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.103117 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.60310276 +0000 UTC m=+147.411672421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.103199 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.106143 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c8r5m" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.188988 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mhc7m" podStartSLOduration=126.188960221 podStartE2EDuration="2m6.188960221s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:41.098799217 +0000 UTC m=+146.907368878" watchObservedRunningTime="2025-11-22 14:29:41.188960221 +0000 UTC m=+146.997529882" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.203582 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.203908 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.703867539 +0000 UTC m=+147.512437200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.204239 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.204632 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.70461901 +0000 UTC m=+147.513188671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.312577 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.313121 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.81310463 +0000 UTC m=+147.621674291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.408759 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:41 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:41 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:41 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.408827 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.416633 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.416952 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:41.916940296 +0000 UTC m=+147.725509957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.471037 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvnps"] Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.471965 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.513578 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.518314 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.518434 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z7nd\" (UniqueName: \"kubernetes.io/projected/8275d661-2656-4e16-bff2-40633bdde977-kube-api-access-6z7nd\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.518508 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-utilities\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.518565 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-catalog-content\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.518653 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.018633401 +0000 UTC m=+147.827203062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.619178 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.619231 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-catalog-content\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.619256 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z7nd\" (UniqueName: \"kubernetes.io/projected/8275d661-2656-4e16-bff2-40633bdde977-kube-api-access-6z7nd\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.619317 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-utilities\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.619681 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-utilities\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.619867 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-catalog-content\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.620225 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.120207552 +0000 UTC m=+147.928777223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.627114 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m54pn"] Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.628298 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.632856 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.660912 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z7nd\" (UniqueName: \"kubernetes.io/projected/8275d661-2656-4e16-bff2-40633bdde977-kube-api-access-6z7nd\") pod \"community-operators-xvnps\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.712648 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m54pn"] Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.722612 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.738952 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvnps"] Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.741571 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.2415431 +0000 UTC m=+148.050112751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.795956 4914 patch_prober.go:28] interesting pod/apiserver-76f77b778f-vhs7k container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]log ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]etcd ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/generic-apiserver-start-informers ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/max-in-flight-filter ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 22 14:29:41 crc kubenswrapper[4914]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 22 14:29:41 crc kubenswrapper[4914]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/project.openshift.io-projectcache ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/openshift.io-startinformers ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 22 14:29:41 crc kubenswrapper[4914]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 22 14:29:41 crc kubenswrapper[4914]: livez check failed Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.796028 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" podUID="3984e8b2-f1ca-48c7-b661-c0f97d7a513c" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.797845 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.825612 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-catalog-content\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.825664 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-utilities\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.825706 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.825741 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crndh\" (UniqueName: \"kubernetes.io/projected/35766d87-0764-410c-ba6c-e0afc70a0379-kube-api-access-crndh\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.826048 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.326036502 +0000 UTC m=+148.134606163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.835079 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ld6hj"] Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.836004 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.855197 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ld6hj"] Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.928375 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.928711 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.428694835 +0000 UTC m=+148.237264496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930172 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9nsh\" (UniqueName: \"kubernetes.io/projected/26230687-39d3-4cfa-ae80-dd176534e764-kube-api-access-w9nsh\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930267 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-catalog-content\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930371 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-utilities\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930464 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930538 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-utilities\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930627 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crndh\" (UniqueName: \"kubernetes.io/projected/35766d87-0764-410c-ba6c-e0afc70a0379-kube-api-access-crndh\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.930712 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-catalog-content\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.931221 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-catalog-content\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.931545 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-utilities\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: E1122 14:29:41.931849 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.431838835 +0000 UTC m=+148.240408496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.964492 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crndh\" (UniqueName: \"kubernetes.io/projected/35766d87-0764-410c-ba6c-e0afc70a0379-kube-api-access-crndh\") pod \"certified-operators-m54pn\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.995974 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mp8qw"] Nov 22 14:29:41 crc kubenswrapper[4914]: I1122 14:29:41.997014 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.011540 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mp8qw"] Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.031506 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.031917 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-catalog-content\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.031956 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9nsh\" (UniqueName: \"kubernetes.io/projected/26230687-39d3-4cfa-ae80-dd176534e764-kube-api-access-w9nsh\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.032022 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-utilities\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.032372 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-utilities\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.032443 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.532427838 +0000 UTC m=+148.340997499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.032643 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-catalog-content\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.055736 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9nsh\" (UniqueName: \"kubernetes.io/projected/26230687-39d3-4cfa-ae80-dd176534e764-kube-api-access-w9nsh\") pod \"community-operators-ld6hj\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.057119 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-25f9c" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.147504 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" event={"ID":"434519f8-3fcf-4a39-9335-9c43f6db4163","Type":"ContainerStarted","Data":"5daace8e3492433d56613890b18d304d5e256f98a2ce4d770509588489f5ada0"} Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.147537 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" event={"ID":"434519f8-3fcf-4a39-9335-9c43f6db4163","Type":"ContainerStarted","Data":"61736d8ac77f943e993716a3c503afcbc4c8a6bfc8445d08d54c937c0685e659"} Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.151300 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-utilities\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.151334 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc7fs\" (UniqueName: \"kubernetes.io/projected/05a64682-151b-4988-b15a-b5647ce67e74-kube-api-access-kc7fs\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.151388 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-catalog-content\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.151466 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.151735 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.651723437 +0000 UTC m=+148.460293088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.183339 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.234932 4914 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.248526 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.253508 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.253675 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.753649329 +0000 UTC m=+148.562218990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.253742 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc7fs\" (UniqueName: \"kubernetes.io/projected/05a64682-151b-4988-b15a-b5647ce67e74-kube-api-access-kc7fs\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.254006 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-catalog-content\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.254273 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.254432 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-utilities\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.255135 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-utilities\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.256201 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.756190531 +0000 UTC m=+148.564760192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.259691 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-catalog-content\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.304387 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc7fs\" (UniqueName: \"kubernetes.io/projected/05a64682-151b-4988-b15a-b5647ce67e74-kube-api-access-kc7fs\") pod \"certified-operators-mp8qw\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.317386 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.339975 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvnps"] Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.359468 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.359900 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.859884334 +0000 UTC m=+148.668453995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: W1122 14:29:42.370034 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8275d661_2656_4e16_bff2_40633bdde977.slice/crio-b17c557d9d658e3e8948dd2ac802c4884d2dfc2ff5d6e854f0ff6479fd8182d1 WatchSource:0}: Error finding container b17c557d9d658e3e8948dd2ac802c4884d2dfc2ff5d6e854f0ff6479fd8182d1: Status 404 returned error can't find the container with id b17c557d9d658e3e8948dd2ac802c4884d2dfc2ff5d6e854f0ff6479fd8182d1 Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.405607 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:42 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:42 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:42 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.405660 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.461886 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.462192 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:42.962180666 +0000 UTC m=+148.770750327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.565672 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.566379 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:43.066364832 +0000 UTC m=+148.874934493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.668908 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.669230 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:43.16921695 +0000 UTC m=+148.977786611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.703202 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ld6hj"] Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.712130 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m54pn"] Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.771128 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.771389 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.771454 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.775595 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:43.275563848 +0000 UTC m=+149.084133509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.776719 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.789639 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.824192 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mp8qw"] Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.873203 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.873890 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.873992 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.875627 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-22 14:29:43.375615216 +0000 UTC m=+149.184184877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mmm9" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.877431 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.878296 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.974777 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:42 crc kubenswrapper[4914]: E1122 14:29:42.975327 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-22 14:29:43.475311384 +0000 UTC m=+149.283881045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 22 14:29:42 crc kubenswrapper[4914]: I1122 14:29:42.988479 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.039034 4914 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-22T14:29:42.234965243Z","Handler":null,"Name":""} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.042004 4914 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.042040 4914 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.076432 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.090575 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.094314 4914 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.094428 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.118000 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.192543 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mmm9\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.226488 4914 generic.go:334] "Generic (PLEG): container finished" podID="26230687-39d3-4cfa-ae80-dd176534e764" containerID="5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe" exitCode=0 Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.226553 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ld6hj" event={"ID":"26230687-39d3-4cfa-ae80-dd176534e764","Type":"ContainerDied","Data":"5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.226581 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ld6hj" event={"ID":"26230687-39d3-4cfa-ae80-dd176534e764","Type":"ContainerStarted","Data":"afc514f011cf646d3af3dd456057cab84b26086b163f65a0a565f1fb5385f2b7"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.239624 4914 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.268582 4914 generic.go:334] "Generic (PLEG): container finished" podID="05a64682-151b-4988-b15a-b5647ce67e74" containerID="13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2" exitCode=0 Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.268655 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp8qw" event={"ID":"05a64682-151b-4988-b15a-b5647ce67e74","Type":"ContainerDied","Data":"13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.268684 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp8qw" event={"ID":"05a64682-151b-4988-b15a-b5647ce67e74","Type":"ContainerStarted","Data":"c43b8ee30f2e634e1211f39a41f1b7530661590692bcec47848ec4fdb0a5830e"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.283432 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.291966 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" event={"ID":"434519f8-3fcf-4a39-9335-9c43f6db4163","Type":"ContainerStarted","Data":"18d1977ac59cb0cd05c35b1d38f4448afc1af54c5dd78c0ced90261af78c3663"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.293651 4914 generic.go:334] "Generic (PLEG): container finished" podID="8275d661-2656-4e16-bff2-40633bdde977" containerID="15c2d4ad9db2b3f91d1322177f4a741389dcfce1ef144eea444628173edb9ba8" exitCode=0 Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.293691 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnps" event={"ID":"8275d661-2656-4e16-bff2-40633bdde977","Type":"ContainerDied","Data":"15c2d4ad9db2b3f91d1322177f4a741389dcfce1ef144eea444628173edb9ba8"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.293706 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnps" event={"ID":"8275d661-2656-4e16-bff2-40633bdde977","Type":"ContainerStarted","Data":"b17c557d9d658e3e8948dd2ac802c4884d2dfc2ff5d6e854f0ff6479fd8182d1"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.301542 4914 generic.go:334] "Generic (PLEG): container finished" podID="35766d87-0764-410c-ba6c-e0afc70a0379" containerID="86136a6f8d1ab7b05945b469084ef591c39cea00f082e37a7eb58ec6a2230679" exitCode=0 Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.302364 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerDied","Data":"86136a6f8d1ab7b05945b469084ef591c39cea00f082e37a7eb58ec6a2230679"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.302383 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerStarted","Data":"1e5fa40e9a58d8765ef22407255be07ac849a21790ae5a5c0f74b3f149c48122"} Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.343293 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.382745 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-nthd7" podStartSLOduration=10.382729722 podStartE2EDuration="10.382729722s" podCreationTimestamp="2025-11-22 14:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:43.380768106 +0000 UTC m=+149.189337767" watchObservedRunningTime="2025-11-22 14:29:43.382729722 +0000 UTC m=+149.191299383" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.385983 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.386711 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.390388 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.390533 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.399571 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qb4hc"] Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.400516 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.401947 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:43 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:43 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:43 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.401979 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.404840 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.415062 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.427995 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qb4hc"] Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.505168 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jlq\" (UniqueName: \"kubernetes.io/projected/f4dd5c74-ad7f-4740-91c1-604264249996-kube-api-access-p2jlq\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.505236 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7391898e-9154-4103-aa0d-ae49ab13fcfb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.505292 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-catalog-content\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.505320 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7391898e-9154-4103-aa0d-ae49ab13fcfb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.505340 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-utilities\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.509303 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.606162 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-catalog-content\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.606557 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7391898e-9154-4103-aa0d-ae49ab13fcfb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.606586 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-utilities\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.606627 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jlq\" (UniqueName: \"kubernetes.io/projected/f4dd5c74-ad7f-4740-91c1-604264249996-kube-api-access-p2jlq\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.606663 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7391898e-9154-4103-aa0d-ae49ab13fcfb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.607645 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-catalog-content\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.607813 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7391898e-9154-4103-aa0d-ae49ab13fcfb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.608097 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-utilities\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.647043 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7391898e-9154-4103-aa0d-ae49ab13fcfb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.647193 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jlq\" (UniqueName: \"kubernetes.io/projected/f4dd5c74-ad7f-4740-91c1-604264249996-kube-api-access-p2jlq\") pod \"redhat-marketplace-qb4hc\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: W1122 14:29:43.663066 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-2d249980db60119a5021caedb2561848e8b2befd438209e30543a9c8bfa0822d WatchSource:0}: Error finding container 2d249980db60119a5021caedb2561848e8b2befd438209e30543a9c8bfa0822d: Status 404 returned error can't find the container with id 2d249980db60119a5021caedb2561848e8b2befd438209e30543a9c8bfa0822d Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.743634 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.805397 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.806282 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r86l9"] Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.807298 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.808992 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r86l9"] Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.913421 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-catalog-content\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.913857 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-utilities\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:43 crc kubenswrapper[4914]: I1122 14:29:43.913962 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvnh9\" (UniqueName: \"kubernetes.io/projected/048fe25e-6b0e-4371-879e-d0b407397aa8-kube-api-access-gvnh9\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.012410 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mmm9"] Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.018341 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvnh9\" (UniqueName: \"kubernetes.io/projected/048fe25e-6b0e-4371-879e-d0b407397aa8-kube-api-access-gvnh9\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.018407 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-catalog-content\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.018430 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-utilities\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.019619 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-catalog-content\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.019856 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-utilities\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.042790 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvnh9\" (UniqueName: \"kubernetes.io/projected/048fe25e-6b0e-4371-879e-d0b407397aa8-kube-api-access-gvnh9\") pod \"redhat-marketplace-r86l9\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.136571 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.139525 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.201761 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qb4hc"] Nov 22 14:29:44 crc kubenswrapper[4914]: W1122 14:29:44.212140 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4dd5c74_ad7f_4740_91c1_604264249996.slice/crio-135517cbc00e1a29ff389b5b88fdbf21b461a57c570ad62736424a9d8b3a8de1 WatchSource:0}: Error finding container 135517cbc00e1a29ff389b5b88fdbf21b461a57c570ad62736424a9d8b3a8de1: Status 404 returned error can't find the container with id 135517cbc00e1a29ff389b5b88fdbf21b461a57c570ad62736424a9d8b3a8de1 Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.319722 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"967275bff55280da896476b580b7d2c665e2c3b92a040ab4d6e02d0ec1ee6d8f"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.320159 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f330c767e001ace0164c7067a439d1067de5f1de53d17a884ad69dd3443bdeed"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.324143 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7391898e-9154-4103-aa0d-ae49ab13fcfb","Type":"ContainerStarted","Data":"6768e4b3ed0279da0340bb694bfe8759e995dabed8a5206adcd247336c5a15c7"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.328680 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e14289e9e44949ccc90be1afbd4a1ccdc85eaca3467ca2813f07d10e15e717d6"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.328706 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2d249980db60119a5021caedb2561848e8b2befd438209e30543a9c8bfa0822d"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.330299 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8f8e23f27d1481838d15853b31ae1ec9867f271890bf86ea669daa164044d247"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.330327 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0d7046272e84b4979c9742db0dc081d057ccdbff3d810174efa2bb350f1d5fd1"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.330513 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.333175 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" event={"ID":"8f152ee7-771c-47ae-bc31-6e4148a017ef","Type":"ContainerStarted","Data":"68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.333204 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" event={"ID":"8f152ee7-771c-47ae-bc31-6e4148a017ef","Type":"ContainerStarted","Data":"a80abd9ff2771448bf48274d29532040624f6ad26f78087d5d020db263edca7d"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.333592 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.338812 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerStarted","Data":"135517cbc00e1a29ff389b5b88fdbf21b461a57c570ad62736424a9d8b3a8de1"} Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.394012 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" podStartSLOduration=129.393996007 podStartE2EDuration="2m9.393996007s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:44.391800445 +0000 UTC m=+150.200370126" watchObservedRunningTime="2025-11-22 14:29:44.393996007 +0000 UTC m=+150.202565668" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.409404 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:44 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:44 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:44 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.409465 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.436595 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r86l9"] Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.590217 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-99q7s"] Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.591335 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.603271 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.606632 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-99q7s"] Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.734100 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-utilities\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.734655 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-catalog-content\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.734697 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85f24\" (UniqueName: \"kubernetes.io/projected/87dd815e-51f0-4214-98f9-2d797a57931a-kube-api-access-85f24\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.838263 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-utilities\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.838346 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-catalog-content\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.838383 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85f24\" (UniqueName: \"kubernetes.io/projected/87dd815e-51f0-4214-98f9-2d797a57931a-kube-api-access-85f24\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.839128 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-utilities\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.839314 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-catalog-content\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.862023 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85f24\" (UniqueName: \"kubernetes.io/projected/87dd815e-51f0-4214-98f9-2d797a57931a-kube-api-access-85f24\") pod \"redhat-operators-99q7s\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.891208 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.961586 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:29:44 crc kubenswrapper[4914]: I1122 14:29:44.986819 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xk4qr"] Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.001882 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xk4qr"] Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.002001 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.146627 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zksbf\" (UniqueName: \"kubernetes.io/projected/05e12f57-26cf-412d-8a06-0aa535536488-kube-api-access-zksbf\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.146693 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-utilities\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.147826 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-catalog-content\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.249978 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-catalog-content\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.250103 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zksbf\" (UniqueName: \"kubernetes.io/projected/05e12f57-26cf-412d-8a06-0aa535536488-kube-api-access-zksbf\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.250148 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-utilities\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.250991 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-utilities\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.256067 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-catalog-content\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.268440 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zksbf\" (UniqueName: \"kubernetes.io/projected/05e12f57-26cf-412d-8a06-0aa535536488-kube-api-access-zksbf\") pod \"redhat-operators-xk4qr\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.269508 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-99q7s"] Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.350218 4914 generic.go:334] "Generic (PLEG): container finished" podID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerID="0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162" exitCode=0 Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.350284 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r86l9" event={"ID":"048fe25e-6b0e-4371-879e-d0b407397aa8","Type":"ContainerDied","Data":"0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162"} Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.350317 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r86l9" event={"ID":"048fe25e-6b0e-4371-879e-d0b407397aa8","Type":"ContainerStarted","Data":"81cafd30a5dfa7f9ac49417ed7418ce6c6044f6d6359574e07e5ca5ae237c95e"} Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.356825 4914 generic.go:334] "Generic (PLEG): container finished" podID="7391898e-9154-4103-aa0d-ae49ab13fcfb" containerID="16a26e7e51a89e76b3491811b3c0a3a4efebb06221560eb6bbb64b9be607fc07" exitCode=0 Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.356929 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7391898e-9154-4103-aa0d-ae49ab13fcfb","Type":"ContainerDied","Data":"16a26e7e51a89e76b3491811b3c0a3a4efebb06221560eb6bbb64b9be607fc07"} Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.361279 4914 generic.go:334] "Generic (PLEG): container finished" podID="f4dd5c74-ad7f-4740-91c1-604264249996" containerID="b273a102f0c2c8cda400566d38bbf3f516fe132d145ee0cc4cb1bead35e93df7" exitCode=0 Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.361368 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerDied","Data":"b273a102f0c2c8cda400566d38bbf3f516fe132d145ee0cc4cb1bead35e93df7"} Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.363607 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.366856 4914 generic.go:334] "Generic (PLEG): container finished" podID="e44ee621-6ddb-4ee6-bf5a-07215d61a280" containerID="e5bfe0155e6c48f5dae83c44bf076f529cfa1827f58e5b7f2300e3a0de019c2b" exitCode=0 Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.367926 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" event={"ID":"e44ee621-6ddb-4ee6-bf5a-07215d61a280","Type":"ContainerDied","Data":"e5bfe0155e6c48f5dae83c44bf076f529cfa1827f58e5b7f2300e3a0de019c2b"} Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.412271 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:45 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:45 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:45 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.412321 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.734493 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xk4qr"] Nov 22 14:29:45 crc kubenswrapper[4914]: W1122 14:29:45.747959 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05e12f57_26cf_412d_8a06_0aa535536488.slice/crio-63792ffbf939e4a9000035140107ce8976c3102d9ee6e1b7780b881b09f2d075 WatchSource:0}: Error finding container 63792ffbf939e4a9000035140107ce8976c3102d9ee6e1b7780b881b09f2d075: Status 404 returned error can't find the container with id 63792ffbf939e4a9000035140107ce8976c3102d9ee6e1b7780b881b09f2d075 Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.775861 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.781712 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-vhs7k" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.808006 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.808064 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.809213 4914 patch_prober.go:28] interesting pod/console-f9d7485db-gsnrn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.809254 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gsnrn" podUID="763d3813-881f-4c0a-8160-5c0a15e17926" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.935665 4914 patch_prober.go:28] interesting pod/downloads-7954f5f757-wshgn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.935710 4914 patch_prober.go:28] interesting pod/downloads-7954f5f757-wshgn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.935716 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wshgn" podUID="033280af-18b4-44cb-bc49-91e77308a685" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 22 14:29:45 crc kubenswrapper[4914]: I1122 14:29:45.935738 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-wshgn" podUID="033280af-18b4-44cb-bc49-91e77308a685" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.401368 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.404952 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 22 14:29:46 crc kubenswrapper[4914]: [-]has-synced failed: reason withheld Nov 22 14:29:46 crc kubenswrapper[4914]: [+]process-running ok Nov 22 14:29:46 crc kubenswrapper[4914]: healthz check failed Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.405051 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.429322 4914 generic.go:334] "Generic (PLEG): container finished" podID="87dd815e-51f0-4214-98f9-2d797a57931a" containerID="d83ff8dc643fc00e70b856970ddbc8b13fc10a0d32871e2c88d452d05d46fa02" exitCode=0 Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.429421 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerDied","Data":"d83ff8dc643fc00e70b856970ddbc8b13fc10a0d32871e2c88d452d05d46fa02"} Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.429449 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerStarted","Data":"b249cba2937a88830c07027e44ea4f32eef0c6bf3bbd304df132012f632d1024"} Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.434023 4914 generic.go:334] "Generic (PLEG): container finished" podID="05e12f57-26cf-412d-8a06-0aa535536488" containerID="dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c" exitCode=0 Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.434989 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerDied","Data":"dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c"} Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.435017 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerStarted","Data":"63792ffbf939e4a9000035140107ce8976c3102d9ee6e1b7780b881b09f2d075"} Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.766594 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.783545 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.795324 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.875089 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqwpc\" (UniqueName: \"kubernetes.io/projected/e44ee621-6ddb-4ee6-bf5a-07215d61a280-kube-api-access-wqwpc\") pod \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.875140 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e44ee621-6ddb-4ee6-bf5a-07215d61a280-config-volume\") pod \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.875184 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7391898e-9154-4103-aa0d-ae49ab13fcfb-kube-api-access\") pod \"7391898e-9154-4103-aa0d-ae49ab13fcfb\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.875201 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7391898e-9154-4103-aa0d-ae49ab13fcfb-kubelet-dir\") pod \"7391898e-9154-4103-aa0d-ae49ab13fcfb\" (UID: \"7391898e-9154-4103-aa0d-ae49ab13fcfb\") " Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.875250 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e44ee621-6ddb-4ee6-bf5a-07215d61a280-secret-volume\") pod \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\" (UID: \"e44ee621-6ddb-4ee6-bf5a-07215d61a280\") " Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.887276 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7391898e-9154-4103-aa0d-ae49ab13fcfb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7391898e-9154-4103-aa0d-ae49ab13fcfb" (UID: "7391898e-9154-4103-aa0d-ae49ab13fcfb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.889433 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e44ee621-6ddb-4ee6-bf5a-07215d61a280-config-volume" (OuterVolumeSpecName: "config-volume") pod "e44ee621-6ddb-4ee6-bf5a-07215d61a280" (UID: "e44ee621-6ddb-4ee6-bf5a-07215d61a280"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.904250 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7391898e-9154-4103-aa0d-ae49ab13fcfb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7391898e-9154-4103-aa0d-ae49ab13fcfb" (UID: "7391898e-9154-4103-aa0d-ae49ab13fcfb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.904514 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e44ee621-6ddb-4ee6-bf5a-07215d61a280-kube-api-access-wqwpc" (OuterVolumeSpecName: "kube-api-access-wqwpc") pod "e44ee621-6ddb-4ee6-bf5a-07215d61a280" (UID: "e44ee621-6ddb-4ee6-bf5a-07215d61a280"). InnerVolumeSpecName "kube-api-access-wqwpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.905231 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e44ee621-6ddb-4ee6-bf5a-07215d61a280-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e44ee621-6ddb-4ee6-bf5a-07215d61a280" (UID: "e44ee621-6ddb-4ee6-bf5a-07215d61a280"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.976708 4914 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e44ee621-6ddb-4ee6-bf5a-07215d61a280-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.976747 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqwpc\" (UniqueName: \"kubernetes.io/projected/e44ee621-6ddb-4ee6-bf5a-07215d61a280-kube-api-access-wqwpc\") on node \"crc\" DevicePath \"\"" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.976756 4914 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e44ee621-6ddb-4ee6-bf5a-07215d61a280-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.976766 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7391898e-9154-4103-aa0d-ae49ab13fcfb-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 14:29:46 crc kubenswrapper[4914]: I1122 14:29:46.976775 4914 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7391898e-9154-4103-aa0d-ae49ab13fcfb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.409845 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.454862 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" event={"ID":"e44ee621-6ddb-4ee6-bf5a-07215d61a280","Type":"ContainerDied","Data":"b385cfbac31c51696b659fd6a4fca4e5c193546e609d2ad3ce1b3187c4677694"} Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.454916 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b385cfbac31c51696b659fd6a4fca4e5c193546e609d2ad3ce1b3187c4677694" Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.454976 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8" Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.463035 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.463154 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7391898e-9154-4103-aa0d-ae49ab13fcfb","Type":"ContainerDied","Data":"6768e4b3ed0279da0340bb694bfe8759e995dabed8a5206adcd247336c5a15c7"} Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.463219 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6768e4b3ed0279da0340bb694bfe8759e995dabed8a5206adcd247336c5a15c7" Nov 22 14:29:47 crc kubenswrapper[4914]: I1122 14:29:47.468126 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-xfwmr" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.024675 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 14:29:49 crc kubenswrapper[4914]: E1122 14:29:49.025266 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e44ee621-6ddb-4ee6-bf5a-07215d61a280" containerName="collect-profiles" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.025280 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e44ee621-6ddb-4ee6-bf5a-07215d61a280" containerName="collect-profiles" Nov 22 14:29:49 crc kubenswrapper[4914]: E1122 14:29:49.025293 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7391898e-9154-4103-aa0d-ae49ab13fcfb" containerName="pruner" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.025299 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="7391898e-9154-4103-aa0d-ae49ab13fcfb" containerName="pruner" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.025420 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="7391898e-9154-4103-aa0d-ae49ab13fcfb" containerName="pruner" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.025446 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="e44ee621-6ddb-4ee6-bf5a-07215d61a280" containerName="collect-profiles" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.025867 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.028447 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.028913 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.029670 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.122589 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53837311-c646-4085-b1ea-1a24fb53f621-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.122723 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53837311-c646-4085-b1ea-1a24fb53f621-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.226219 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53837311-c646-4085-b1ea-1a24fb53f621-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.226320 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53837311-c646-4085-b1ea-1a24fb53f621-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.226685 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53837311-c646-4085-b1ea-1a24fb53f621-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.246084 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53837311-c646-4085-b1ea-1a24fb53f621-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:49 crc kubenswrapper[4914]: I1122 14:29:49.370836 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:29:50 crc kubenswrapper[4914]: I1122 14:29:50.057784 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 22 14:29:50 crc kubenswrapper[4914]: I1122 14:29:50.505285 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53837311-c646-4085-b1ea-1a24fb53f621","Type":"ContainerStarted","Data":"675f9b9bd731ce2bf0dab83d514c184285f25db636f9b68738bff1ca21c4b0cd"} Nov 22 14:29:50 crc kubenswrapper[4914]: I1122 14:29:50.946893 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:29:51 crc kubenswrapper[4914]: I1122 14:29:51.522352 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53837311-c646-4085-b1ea-1a24fb53f621","Type":"ContainerStarted","Data":"10ccdad85c9b50eec0a3ad54b1df6a754646d47f0d057df353a1bca49c503ab4"} Nov 22 14:29:51 crc kubenswrapper[4914]: I1122 14:29:51.538674 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.538658347 podStartE2EDuration="3.538658347s" podCreationTimestamp="2025-11-22 14:29:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:29:51.53460512 +0000 UTC m=+157.343174791" watchObservedRunningTime="2025-11-22 14:29:51.538658347 +0000 UTC m=+157.347228008" Nov 22 14:29:51 crc kubenswrapper[4914]: I1122 14:29:51.851645 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gwt8v" Nov 22 14:29:52 crc kubenswrapper[4914]: I1122 14:29:52.046478 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:29:52 crc kubenswrapper[4914]: I1122 14:29:52.047320 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:29:52 crc kubenswrapper[4914]: I1122 14:29:52.533055 4914 generic.go:334] "Generic (PLEG): container finished" podID="53837311-c646-4085-b1ea-1a24fb53f621" containerID="10ccdad85c9b50eec0a3ad54b1df6a754646d47f0d057df353a1bca49c503ab4" exitCode=0 Nov 22 14:29:52 crc kubenswrapper[4914]: I1122 14:29:52.533370 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53837311-c646-4085-b1ea-1a24fb53f621","Type":"ContainerDied","Data":"10ccdad85c9b50eec0a3ad54b1df6a754646d47f0d057df353a1bca49c503ab4"} Nov 22 14:29:55 crc kubenswrapper[4914]: I1122 14:29:55.818989 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:55 crc kubenswrapper[4914]: I1122 14:29:55.822762 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-gsnrn" Nov 22 14:29:55 crc kubenswrapper[4914]: I1122 14:29:55.949051 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-wshgn" Nov 22 14:29:57 crc kubenswrapper[4914]: I1122 14:29:57.776164 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:57 crc kubenswrapper[4914]: I1122 14:29:57.784551 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d8b0b491-074a-465a-9419-6481d7877ecf-metrics-certs\") pod \"network-metrics-daemon-47bbd\" (UID: \"d8b0b491-074a-465a-9419-6481d7877ecf\") " pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:29:57 crc kubenswrapper[4914]: I1122 14:29:57.805518 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-47bbd" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.148856 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz"] Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.149715 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.152535 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.152727 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.166497 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz"] Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.312303 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0802ab8f-68e3-41fa-b454-cbbdedd10f29-secret-volume\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.312393 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0802ab8f-68e3-41fa-b454-cbbdedd10f29-config-volume\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.312443 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnc7z\" (UniqueName: \"kubernetes.io/projected/0802ab8f-68e3-41fa-b454-cbbdedd10f29-kube-api-access-wnc7z\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.414074 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnc7z\" (UniqueName: \"kubernetes.io/projected/0802ab8f-68e3-41fa-b454-cbbdedd10f29-kube-api-access-wnc7z\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.414195 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0802ab8f-68e3-41fa-b454-cbbdedd10f29-secret-volume\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.414237 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0802ab8f-68e3-41fa-b454-cbbdedd10f29-config-volume\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.415292 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0802ab8f-68e3-41fa-b454-cbbdedd10f29-config-volume\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.424954 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0802ab8f-68e3-41fa-b454-cbbdedd10f29-secret-volume\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.443688 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnc7z\" (UniqueName: \"kubernetes.io/projected/0802ab8f-68e3-41fa-b454-cbbdedd10f29-kube-api-access-wnc7z\") pod \"collect-profiles-29397030-gjdvz\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.476065 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.505806 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.598347 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53837311-c646-4085-b1ea-1a24fb53f621","Type":"ContainerDied","Data":"675f9b9bd731ce2bf0dab83d514c184285f25db636f9b68738bff1ca21c4b0cd"} Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.598388 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="675f9b9bd731ce2bf0dab83d514c184285f25db636f9b68738bff1ca21c4b0cd" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.599001 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.617073 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53837311-c646-4085-b1ea-1a24fb53f621-kube-api-access\") pod \"53837311-c646-4085-b1ea-1a24fb53f621\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.617253 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53837311-c646-4085-b1ea-1a24fb53f621-kubelet-dir\") pod \"53837311-c646-4085-b1ea-1a24fb53f621\" (UID: \"53837311-c646-4085-b1ea-1a24fb53f621\") " Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.617528 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53837311-c646-4085-b1ea-1a24fb53f621-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "53837311-c646-4085-b1ea-1a24fb53f621" (UID: "53837311-c646-4085-b1ea-1a24fb53f621"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.621518 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53837311-c646-4085-b1ea-1a24fb53f621-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "53837311-c646-4085-b1ea-1a24fb53f621" (UID: "53837311-c646-4085-b1ea-1a24fb53f621"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.718792 4914 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53837311-c646-4085-b1ea-1a24fb53f621-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:00 crc kubenswrapper[4914]: I1122 14:30:00.718831 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53837311-c646-4085-b1ea-1a24fb53f621-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:03 crc kubenswrapper[4914]: I1122 14:30:03.516349 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:30:07 crc kubenswrapper[4914]: I1122 14:30:07.439095 4914 patch_prober.go:28] interesting pod/router-default-5444994796-xfwmr container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 22 14:30:07 crc kubenswrapper[4914]: I1122 14:30:07.439398 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-xfwmr" podUID="baad2152-3fdb-4a99-baee-17471ae8dfc5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 22 14:30:16 crc kubenswrapper[4914]: I1122 14:30:16.079068 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d5bhj" Nov 22 14:30:17 crc kubenswrapper[4914]: E1122 14:30:17.537715 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 22 14:30:17 crc kubenswrapper[4914]: E1122 14:30:17.538187 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2jlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qb4hc_openshift-marketplace(f4dd5c74-ad7f-4740-91c1-604264249996): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:17 crc kubenswrapper[4914]: E1122 14:30:17.539342 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-qb4hc" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" Nov 22 14:30:18 crc kubenswrapper[4914]: E1122 14:30:18.734931 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qb4hc" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" Nov 22 14:30:18 crc kubenswrapper[4914]: E1122 14:30:18.791290 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 14:30:18 crc kubenswrapper[4914]: E1122 14:30:18.791395 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6z7nd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xvnps_openshift-marketplace(8275d661-2656-4e16-bff2-40633bdde977): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:18 crc kubenswrapper[4914]: E1122 14:30:18.793608 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xvnps" podUID="8275d661-2656-4e16-bff2-40633bdde977" Nov 22 14:30:20 crc kubenswrapper[4914]: E1122 14:30:20.144151 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xvnps" podUID="8275d661-2656-4e16-bff2-40633bdde977" Nov 22 14:30:20 crc kubenswrapper[4914]: E1122 14:30:20.256487 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 14:30:20 crc kubenswrapper[4914]: E1122 14:30:20.256722 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crndh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-m54pn_openshift-marketplace(35766d87-0764-410c-ba6c-e0afc70a0379): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:20 crc kubenswrapper[4914]: E1122 14:30:20.257985 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-m54pn" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" Nov 22 14:30:22 crc kubenswrapper[4914]: I1122 14:30:22.042867 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:30:22 crc kubenswrapper[4914]: I1122 14:30:22.042947 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.637623 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-m54pn" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.717173 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.717331 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-85f24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-99q7s_openshift-marketplace(87dd815e-51f0-4214-98f9-2d797a57931a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.718496 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-99q7s" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.739783 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-99q7s" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.744906 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.745021 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w9nsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ld6hj_openshift-marketplace(26230687-39d3-4cfa-ae80-dd176534e764): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.746282 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ld6hj" podUID="26230687-39d3-4cfa-ae80-dd176534e764" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.769624 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.769783 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zksbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xk4qr_openshift-marketplace(05e12f57-26cf-412d-8a06-0aa535536488): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.771690 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xk4qr" podUID="05e12f57-26cf-412d-8a06-0aa535536488" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.797360 4914 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.797548 4914 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kc7fs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mp8qw_openshift-marketplace(05a64682-151b-4988-b15a-b5647ce67e74): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 22 14:30:22 crc kubenswrapper[4914]: E1122 14:30:22.798715 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mp8qw" podUID="05a64682-151b-4988-b15a-b5647ce67e74" Nov 22 14:30:22 crc kubenswrapper[4914]: I1122 14:30:22.881790 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz"] Nov 22 14:30:22 crc kubenswrapper[4914]: W1122 14:30:22.892561 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0802ab8f_68e3_41fa_b454_cbbdedd10f29.slice/crio-f22dc861f43f154dc55220f96591770511806c1a69a92c320a51a9e2af976ed5 WatchSource:0}: Error finding container f22dc861f43f154dc55220f96591770511806c1a69a92c320a51a9e2af976ed5: Status 404 returned error can't find the container with id f22dc861f43f154dc55220f96591770511806c1a69a92c320a51a9e2af976ed5 Nov 22 14:30:22 crc kubenswrapper[4914]: I1122 14:30:22.917395 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-47bbd"] Nov 22 14:30:22 crc kubenswrapper[4914]: W1122 14:30:22.920232 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8b0b491_074a_465a_9419_6481d7877ecf.slice/crio-7d492897d52d290aefd312f68ee6e5a3d075284070f935203aa98f25324f6656 WatchSource:0}: Error finding container 7d492897d52d290aefd312f68ee6e5a3d075284070f935203aa98f25324f6656: Status 404 returned error can't find the container with id 7d492897d52d290aefd312f68ee6e5a3d075284070f935203aa98f25324f6656 Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.123254 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.746271 4914 generic.go:334] "Generic (PLEG): container finished" podID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerID="1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c" exitCode=0 Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.746347 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r86l9" event={"ID":"048fe25e-6b0e-4371-879e-d0b407397aa8","Type":"ContainerDied","Data":"1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c"} Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.753404 4914 generic.go:334] "Generic (PLEG): container finished" podID="0802ab8f-68e3-41fa-b454-cbbdedd10f29" containerID="5ada5366dc3892fa4b8c1db6811376c4ce919a6a1330475c8a00c00a7cf9823f" exitCode=0 Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.753543 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" event={"ID":"0802ab8f-68e3-41fa-b454-cbbdedd10f29","Type":"ContainerDied","Data":"5ada5366dc3892fa4b8c1db6811376c4ce919a6a1330475c8a00c00a7cf9823f"} Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.753584 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" event={"ID":"0802ab8f-68e3-41fa-b454-cbbdedd10f29","Type":"ContainerStarted","Data":"f22dc861f43f154dc55220f96591770511806c1a69a92c320a51a9e2af976ed5"} Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.756928 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-47bbd" event={"ID":"d8b0b491-074a-465a-9419-6481d7877ecf","Type":"ContainerStarted","Data":"24756143b267cc09972ce4a033685f42f851f528421824ce929f745cdea31416"} Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.756985 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-47bbd" event={"ID":"d8b0b491-074a-465a-9419-6481d7877ecf","Type":"ContainerStarted","Data":"a7647c0475b8524c9c8ce5a7da4fb8554bd381dc83367bb6ad5c3c9fa4634e2f"} Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.757007 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-47bbd" event={"ID":"d8b0b491-074a-465a-9419-6481d7877ecf","Type":"ContainerStarted","Data":"7d492897d52d290aefd312f68ee6e5a3d075284070f935203aa98f25324f6656"} Nov 22 14:30:23 crc kubenswrapper[4914]: E1122 14:30:23.760358 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mp8qw" podUID="05a64682-151b-4988-b15a-b5647ce67e74" Nov 22 14:30:23 crc kubenswrapper[4914]: E1122 14:30:23.760629 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ld6hj" podUID="26230687-39d3-4cfa-ae80-dd176534e764" Nov 22 14:30:23 crc kubenswrapper[4914]: E1122 14:30:23.762199 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xk4qr" podUID="05e12f57-26cf-412d-8a06-0aa535536488" Nov 22 14:30:23 crc kubenswrapper[4914]: I1122 14:30:23.853797 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-47bbd" podStartSLOduration=168.853781074 podStartE2EDuration="2m48.853781074s" podCreationTimestamp="2025-11-22 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:30:23.851981553 +0000 UTC m=+189.660551254" watchObservedRunningTime="2025-11-22 14:30:23.853781074 +0000 UTC m=+189.662350725" Nov 22 14:30:24 crc kubenswrapper[4914]: I1122 14:30:24.766592 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r86l9" event={"ID":"048fe25e-6b0e-4371-879e-d0b407397aa8","Type":"ContainerStarted","Data":"d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9"} Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.049618 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.065715 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r86l9" podStartSLOduration=3.25780607 podStartE2EDuration="42.065705111s" podCreationTimestamp="2025-11-22 14:29:43 +0000 UTC" firstStartedPulling="2025-11-22 14:29:45.359838932 +0000 UTC m=+151.168408593" lastFinishedPulling="2025-11-22 14:30:24.167737973 +0000 UTC m=+189.976307634" observedRunningTime="2025-11-22 14:30:24.79558636 +0000 UTC m=+190.604156031" watchObservedRunningTime="2025-11-22 14:30:25.065705111 +0000 UTC m=+190.874274772" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.205536 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnc7z\" (UniqueName: \"kubernetes.io/projected/0802ab8f-68e3-41fa-b454-cbbdedd10f29-kube-api-access-wnc7z\") pod \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.205928 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0802ab8f-68e3-41fa-b454-cbbdedd10f29-secret-volume\") pod \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.206008 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0802ab8f-68e3-41fa-b454-cbbdedd10f29-config-volume\") pod \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\" (UID: \"0802ab8f-68e3-41fa-b454-cbbdedd10f29\") " Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.206737 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0802ab8f-68e3-41fa-b454-cbbdedd10f29-config-volume" (OuterVolumeSpecName: "config-volume") pod "0802ab8f-68e3-41fa-b454-cbbdedd10f29" (UID: "0802ab8f-68e3-41fa-b454-cbbdedd10f29"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.215011 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0802ab8f-68e3-41fa-b454-cbbdedd10f29-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0802ab8f-68e3-41fa-b454-cbbdedd10f29" (UID: "0802ab8f-68e3-41fa-b454-cbbdedd10f29"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.222199 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0802ab8f-68e3-41fa-b454-cbbdedd10f29-kube-api-access-wnc7z" (OuterVolumeSpecName: "kube-api-access-wnc7z") pod "0802ab8f-68e3-41fa-b454-cbbdedd10f29" (UID: "0802ab8f-68e3-41fa-b454-cbbdedd10f29"). InnerVolumeSpecName "kube-api-access-wnc7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.307389 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnc7z\" (UniqueName: \"kubernetes.io/projected/0802ab8f-68e3-41fa-b454-cbbdedd10f29-kube-api-access-wnc7z\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.307433 4914 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0802ab8f-68e3-41fa-b454-cbbdedd10f29-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.307446 4914 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0802ab8f-68e3-41fa-b454-cbbdedd10f29-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.772731 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.773262 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397030-gjdvz" event={"ID":"0802ab8f-68e3-41fa-b454-cbbdedd10f29","Type":"ContainerDied","Data":"f22dc861f43f154dc55220f96591770511806c1a69a92c320a51a9e2af976ed5"} Nov 22 14:30:25 crc kubenswrapper[4914]: I1122 14:30:25.773295 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f22dc861f43f154dc55220f96591770511806c1a69a92c320a51a9e2af976ed5" Nov 22 14:30:32 crc kubenswrapper[4914]: I1122 14:30:32.818061 4914 generic.go:334] "Generic (PLEG): container finished" podID="8275d661-2656-4e16-bff2-40633bdde977" containerID="5366e763a2591e5a36195e4674dd314f69a78215fb134a19545a0375700c3778" exitCode=0 Nov 22 14:30:32 crc kubenswrapper[4914]: I1122 14:30:32.818119 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnps" event={"ID":"8275d661-2656-4e16-bff2-40633bdde977","Type":"ContainerDied","Data":"5366e763a2591e5a36195e4674dd314f69a78215fb134a19545a0375700c3778"} Nov 22 14:30:33 crc kubenswrapper[4914]: I1122 14:30:33.858670 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnps" event={"ID":"8275d661-2656-4e16-bff2-40633bdde977","Type":"ContainerStarted","Data":"cf86376eeb5c3a7e88bfa303595d94c75a66a6c1796ea31feb52793efc756995"} Nov 22 14:30:33 crc kubenswrapper[4914]: I1122 14:30:33.873620 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvnps" podStartSLOduration=2.806292071 podStartE2EDuration="52.87360607s" podCreationTimestamp="2025-11-22 14:29:41 +0000 UTC" firstStartedPulling="2025-11-22 14:29:43.29998804 +0000 UTC m=+149.108557701" lastFinishedPulling="2025-11-22 14:30:33.367302039 +0000 UTC m=+199.175871700" observedRunningTime="2025-11-22 14:30:33.873256942 +0000 UTC m=+199.681826603" watchObservedRunningTime="2025-11-22 14:30:33.87360607 +0000 UTC m=+199.682175721" Nov 22 14:30:34 crc kubenswrapper[4914]: I1122 14:30:34.137498 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:30:34 crc kubenswrapper[4914]: I1122 14:30:34.137563 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:30:34 crc kubenswrapper[4914]: I1122 14:30:34.255946 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:30:34 crc kubenswrapper[4914]: I1122 14:30:34.874655 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerStarted","Data":"8ee2c0dce00bba6aeb82e1470ef8540119ddd09466828bcd98ef81d65fcd23f9"} Nov 22 14:30:34 crc kubenswrapper[4914]: I1122 14:30:34.906972 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:30:35 crc kubenswrapper[4914]: I1122 14:30:35.873824 4914 generic.go:334] "Generic (PLEG): container finished" podID="f4dd5c74-ad7f-4740-91c1-604264249996" containerID="8ee2c0dce00bba6aeb82e1470ef8540119ddd09466828bcd98ef81d65fcd23f9" exitCode=0 Nov 22 14:30:35 crc kubenswrapper[4914]: I1122 14:30:35.873920 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerDied","Data":"8ee2c0dce00bba6aeb82e1470ef8540119ddd09466828bcd98ef81d65fcd23f9"} Nov 22 14:30:36 crc kubenswrapper[4914]: I1122 14:30:36.082492 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r86l9"] Nov 22 14:30:36 crc kubenswrapper[4914]: I1122 14:30:36.899307 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r86l9" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="registry-server" containerID="cri-o://d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9" gracePeriod=2 Nov 22 14:30:36 crc kubenswrapper[4914]: I1122 14:30:36.899598 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerStarted","Data":"fca1a112a12969e6881ba69fc04107601a4fa37a917944d1cd95c19b0b3a80f7"} Nov 22 14:30:36 crc kubenswrapper[4914]: I1122 14:30:36.899727 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerStarted","Data":"7cdb66731fcb66fb1d2f7e1dfc3483f91cde0d2405fcece30f028c9dfe8aa233"} Nov 22 14:30:36 crc kubenswrapper[4914]: I1122 14:30:36.963023 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qb4hc" podStartSLOduration=3.087034064 podStartE2EDuration="53.963004885s" podCreationTimestamp="2025-11-22 14:29:43 +0000 UTC" firstStartedPulling="2025-11-22 14:29:45.390636355 +0000 UTC m=+151.199206016" lastFinishedPulling="2025-11-22 14:30:36.266607176 +0000 UTC m=+202.075176837" observedRunningTime="2025-11-22 14:30:36.960160371 +0000 UTC m=+202.768730042" watchObservedRunningTime="2025-11-22 14:30:36.963004885 +0000 UTC m=+202.771574546" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.238560 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.354440 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-utilities\") pod \"048fe25e-6b0e-4371-879e-d0b407397aa8\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.354757 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvnh9\" (UniqueName: \"kubernetes.io/projected/048fe25e-6b0e-4371-879e-d0b407397aa8-kube-api-access-gvnh9\") pod \"048fe25e-6b0e-4371-879e-d0b407397aa8\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.354794 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-catalog-content\") pod \"048fe25e-6b0e-4371-879e-d0b407397aa8\" (UID: \"048fe25e-6b0e-4371-879e-d0b407397aa8\") " Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.355277 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-utilities" (OuterVolumeSpecName: "utilities") pod "048fe25e-6b0e-4371-879e-d0b407397aa8" (UID: "048fe25e-6b0e-4371-879e-d0b407397aa8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.360298 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048fe25e-6b0e-4371-879e-d0b407397aa8-kube-api-access-gvnh9" (OuterVolumeSpecName: "kube-api-access-gvnh9") pod "048fe25e-6b0e-4371-879e-d0b407397aa8" (UID: "048fe25e-6b0e-4371-879e-d0b407397aa8"). InnerVolumeSpecName "kube-api-access-gvnh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.374303 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "048fe25e-6b0e-4371-879e-d0b407397aa8" (UID: "048fe25e-6b0e-4371-879e-d0b407397aa8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.455847 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvnh9\" (UniqueName: \"kubernetes.io/projected/048fe25e-6b0e-4371-879e-d0b407397aa8-kube-api-access-gvnh9\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.455906 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.455920 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048fe25e-6b0e-4371-879e-d0b407397aa8-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.912958 4914 generic.go:334] "Generic (PLEG): container finished" podID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerID="d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9" exitCode=0 Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.913008 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r86l9" event={"ID":"048fe25e-6b0e-4371-879e-d0b407397aa8","Type":"ContainerDied","Data":"d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9"} Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.913068 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r86l9" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.913090 4914 scope.go:117] "RemoveContainer" containerID="d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.913074 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r86l9" event={"ID":"048fe25e-6b0e-4371-879e-d0b407397aa8","Type":"ContainerDied","Data":"81cafd30a5dfa7f9ac49417ed7418ce6c6044f6d6359574e07e5ca5ae237c95e"} Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.919026 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerStarted","Data":"938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614"} Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.921976 4914 generic.go:334] "Generic (PLEG): container finished" podID="26230687-39d3-4cfa-ae80-dd176534e764" containerID="8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde" exitCode=0 Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.922054 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ld6hj" event={"ID":"26230687-39d3-4cfa-ae80-dd176534e764","Type":"ContainerDied","Data":"8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde"} Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.927013 4914 generic.go:334] "Generic (PLEG): container finished" podID="05a64682-151b-4988-b15a-b5647ce67e74" containerID="85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4" exitCode=0 Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.927079 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp8qw" event={"ID":"05a64682-151b-4988-b15a-b5647ce67e74","Type":"ContainerDied","Data":"85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4"} Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.928953 4914 generic.go:334] "Generic (PLEG): container finished" podID="87dd815e-51f0-4214-98f9-2d797a57931a" containerID="7cdb66731fcb66fb1d2f7e1dfc3483f91cde0d2405fcece30f028c9dfe8aa233" exitCode=0 Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.928984 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerDied","Data":"7cdb66731fcb66fb1d2f7e1dfc3483f91cde0d2405fcece30f028c9dfe8aa233"} Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.946741 4914 scope.go:117] "RemoveContainer" containerID="1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.961453 4914 scope.go:117] "RemoveContainer" containerID="0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.984314 4914 scope.go:117] "RemoveContainer" containerID="d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9" Nov 22 14:30:37 crc kubenswrapper[4914]: E1122 14:30:37.984909 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9\": container with ID starting with d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9 not found: ID does not exist" containerID="d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.984952 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9"} err="failed to get container status \"d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9\": rpc error: code = NotFound desc = could not find container \"d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9\": container with ID starting with d81a0106baf595f4da2ac5ee1f6273e4891a92415c6ecbdb9e66fd52db2018a9 not found: ID does not exist" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.985026 4914 scope.go:117] "RemoveContainer" containerID="1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c" Nov 22 14:30:37 crc kubenswrapper[4914]: E1122 14:30:37.986464 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c\": container with ID starting with 1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c not found: ID does not exist" containerID="1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.986497 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c"} err="failed to get container status \"1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c\": rpc error: code = NotFound desc = could not find container \"1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c\": container with ID starting with 1e8e5dbf92202456548096850f677b504afea40bac95d66c6439bca12219fd0c not found: ID does not exist" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.986521 4914 scope.go:117] "RemoveContainer" containerID="0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162" Nov 22 14:30:37 crc kubenswrapper[4914]: E1122 14:30:37.988311 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162\": container with ID starting with 0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162 not found: ID does not exist" containerID="0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162" Nov 22 14:30:37 crc kubenswrapper[4914]: I1122 14:30:37.988342 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162"} err="failed to get container status \"0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162\": rpc error: code = NotFound desc = could not find container \"0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162\": container with ID starting with 0f2ae7e8bd0bc69015224dfa40111516e898d8f9a90a9f73721b8914a5b3b162 not found: ID does not exist" Nov 22 14:30:38 crc kubenswrapper[4914]: I1122 14:30:38.005559 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r86l9"] Nov 22 14:30:38 crc kubenswrapper[4914]: I1122 14:30:38.008070 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r86l9"] Nov 22 14:30:38 crc kubenswrapper[4914]: I1122 14:30:38.878051 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" path="/var/lib/kubelet/pods/048fe25e-6b0e-4371-879e-d0b407397aa8/volumes" Nov 22 14:30:38 crc kubenswrapper[4914]: I1122 14:30:38.935680 4914 generic.go:334] "Generic (PLEG): container finished" podID="05e12f57-26cf-412d-8a06-0aa535536488" containerID="938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614" exitCode=0 Nov 22 14:30:38 crc kubenswrapper[4914]: I1122 14:30:38.935759 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerDied","Data":"938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614"} Nov 22 14:30:39 crc kubenswrapper[4914]: I1122 14:30:39.958954 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerStarted","Data":"f4285cff2fb23b58c1f5c16573f8a8a6b38e183919c8e79edd6a755b681b943f"} Nov 22 14:30:39 crc kubenswrapper[4914]: I1122 14:30:39.981536 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-99q7s" podStartSLOduration=3.538874553 podStartE2EDuration="55.981519574s" podCreationTimestamp="2025-11-22 14:29:44 +0000 UTC" firstStartedPulling="2025-11-22 14:29:46.436576205 +0000 UTC m=+152.245145866" lastFinishedPulling="2025-11-22 14:30:38.879221226 +0000 UTC m=+204.687790887" observedRunningTime="2025-11-22 14:30:39.976139844 +0000 UTC m=+205.784709505" watchObservedRunningTime="2025-11-22 14:30:39.981519574 +0000 UTC m=+205.790089235" Nov 22 14:30:41 crc kubenswrapper[4914]: I1122 14:30:41.798727 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:30:41 crc kubenswrapper[4914]: I1122 14:30:41.799148 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:30:41 crc kubenswrapper[4914]: I1122 14:30:41.847295 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:30:41 crc kubenswrapper[4914]: I1122 14:30:41.970648 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp8qw" event={"ID":"05a64682-151b-4988-b15a-b5647ce67e74","Type":"ContainerStarted","Data":"06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173"} Nov 22 14:30:41 crc kubenswrapper[4914]: I1122 14:30:41.995894 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mp8qw" podStartSLOduration=4.629676289 podStartE2EDuration="1m0.995874429s" podCreationTimestamp="2025-11-22 14:29:41 +0000 UTC" firstStartedPulling="2025-11-22 14:29:43.284522127 +0000 UTC m=+149.093091788" lastFinishedPulling="2025-11-22 14:30:39.650720267 +0000 UTC m=+205.459289928" observedRunningTime="2025-11-22 14:30:41.993890617 +0000 UTC m=+207.802460278" watchObservedRunningTime="2025-11-22 14:30:41.995874429 +0000 UTC m=+207.804444090" Nov 22 14:30:42 crc kubenswrapper[4914]: I1122 14:30:42.010440 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:30:42 crc kubenswrapper[4914]: I1122 14:30:42.317944 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:30:42 crc kubenswrapper[4914]: I1122 14:30:42.318002 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:30:43 crc kubenswrapper[4914]: I1122 14:30:43.365525 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mp8qw" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="registry-server" probeResult="failure" output=< Nov 22 14:30:43 crc kubenswrapper[4914]: timeout: failed to connect service ":50051" within 1s Nov 22 14:30:43 crc kubenswrapper[4914]: > Nov 22 14:30:43 crc kubenswrapper[4914]: I1122 14:30:43.807407 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:30:43 crc kubenswrapper[4914]: I1122 14:30:43.808001 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:30:43 crc kubenswrapper[4914]: I1122 14:30:43.871698 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:30:44 crc kubenswrapper[4914]: I1122 14:30:44.023632 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:30:44 crc kubenswrapper[4914]: I1122 14:30:44.962672 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:30:44 crc kubenswrapper[4914]: I1122 14:30:44.963167 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:30:45 crc kubenswrapper[4914]: I1122 14:30:45.998372 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-99q7s" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="registry-server" probeResult="failure" output=< Nov 22 14:30:45 crc kubenswrapper[4914]: timeout: failed to connect service ":50051" within 1s Nov 22 14:30:45 crc kubenswrapper[4914]: > Nov 22 14:30:46 crc kubenswrapper[4914]: I1122 14:30:46.001268 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerStarted","Data":"08c3c465ed9e6e79713084ebbfd77f13ad671d644873efbece6e91a602657982"} Nov 22 14:30:46 crc kubenswrapper[4914]: I1122 14:30:46.003611 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ld6hj" event={"ID":"26230687-39d3-4cfa-ae80-dd176534e764","Type":"ContainerStarted","Data":"82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7"} Nov 22 14:30:47 crc kubenswrapper[4914]: I1122 14:30:47.016713 4914 generic.go:334] "Generic (PLEG): container finished" podID="35766d87-0764-410c-ba6c-e0afc70a0379" containerID="08c3c465ed9e6e79713084ebbfd77f13ad671d644873efbece6e91a602657982" exitCode=0 Nov 22 14:30:47 crc kubenswrapper[4914]: I1122 14:30:47.016833 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerDied","Data":"08c3c465ed9e6e79713084ebbfd77f13ad671d644873efbece6e91a602657982"} Nov 22 14:30:47 crc kubenswrapper[4914]: I1122 14:30:47.038170 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ld6hj" podStartSLOduration=4.849676642 podStartE2EDuration="1m6.038145357s" podCreationTimestamp="2025-11-22 14:29:41 +0000 UTC" firstStartedPulling="2025-11-22 14:29:43.239400214 +0000 UTC m=+149.047969875" lastFinishedPulling="2025-11-22 14:30:44.427868929 +0000 UTC m=+210.236438590" observedRunningTime="2025-11-22 14:30:46.029490677 +0000 UTC m=+211.838060348" watchObservedRunningTime="2025-11-22 14:30:47.038145357 +0000 UTC m=+212.846715058" Nov 22 14:30:49 crc kubenswrapper[4914]: I1122 14:30:49.029516 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerStarted","Data":"443a98405c3fdea9a0192100b4ed00005cb292baa11971410303508953c88e8f"} Nov 22 14:30:49 crc kubenswrapper[4914]: I1122 14:30:49.033206 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerStarted","Data":"f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51"} Nov 22 14:30:49 crc kubenswrapper[4914]: I1122 14:30:49.051826 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m54pn" podStartSLOduration=2.554624154 podStartE2EDuration="1m8.051810616s" podCreationTimestamp="2025-11-22 14:29:41 +0000 UTC" firstStartedPulling="2025-11-22 14:29:43.326195621 +0000 UTC m=+149.134765312" lastFinishedPulling="2025-11-22 14:30:48.823382113 +0000 UTC m=+214.631951774" observedRunningTime="2025-11-22 14:30:49.049348861 +0000 UTC m=+214.857918532" watchObservedRunningTime="2025-11-22 14:30:49.051810616 +0000 UTC m=+214.860380277" Nov 22 14:30:49 crc kubenswrapper[4914]: I1122 14:30:49.063351 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xk4qr" podStartSLOduration=3.780481889 podStartE2EDuration="1m5.063333405s" podCreationTimestamp="2025-11-22 14:29:44 +0000 UTC" firstStartedPulling="2025-11-22 14:29:46.436828072 +0000 UTC m=+152.245397733" lastFinishedPulling="2025-11-22 14:30:47.719679588 +0000 UTC m=+213.528249249" observedRunningTime="2025-11-22 14:30:49.062351789 +0000 UTC m=+214.870921470" watchObservedRunningTime="2025-11-22 14:30:49.063333405 +0000 UTC m=+214.871903066" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.042388 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.042997 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.043176 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.044253 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.044392 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48" gracePeriod=600 Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.183962 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.184034 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.231363 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.249464 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.249516 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.314350 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.365401 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:30:52 crc kubenswrapper[4914]: I1122 14:30:52.412628 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:30:53 crc kubenswrapper[4914]: I1122 14:30:53.062231 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48" exitCode=0 Nov 22 14:30:53 crc kubenswrapper[4914]: I1122 14:30:53.062354 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48"} Nov 22 14:30:53 crc kubenswrapper[4914]: I1122 14:30:53.062670 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"295e4be82247193043ad744eafabe554329435763a26422ff3518c50082e23e2"} Nov 22 14:30:53 crc kubenswrapper[4914]: I1122 14:30:53.127366 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:30:53 crc kubenswrapper[4914]: I1122 14:30:53.147588 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:30:54 crc kubenswrapper[4914]: I1122 14:30:54.081530 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ld6hj"] Nov 22 14:30:54 crc kubenswrapper[4914]: I1122 14:30:54.683238 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mp8qw"] Nov 22 14:30:54 crc kubenswrapper[4914]: I1122 14:30:54.683822 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mp8qw" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="registry-server" containerID="cri-o://06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173" gracePeriod=2 Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.005839 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.024465 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.051448 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.078070 4914 generic.go:334] "Generic (PLEG): container finished" podID="05a64682-151b-4988-b15a-b5647ce67e74" containerID="06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173" exitCode=0 Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.078293 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ld6hj" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="registry-server" containerID="cri-o://82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7" gracePeriod=2 Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.078579 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp8qw" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.078583 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp8qw" event={"ID":"05a64682-151b-4988-b15a-b5647ce67e74","Type":"ContainerDied","Data":"06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173"} Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.078634 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp8qw" event={"ID":"05a64682-151b-4988-b15a-b5647ce67e74","Type":"ContainerDied","Data":"c43b8ee30f2e634e1211f39a41f1b7530661590692bcec47848ec4fdb0a5830e"} Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.078653 4914 scope.go:117] "RemoveContainer" containerID="06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.095066 4914 scope.go:117] "RemoveContainer" containerID="85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.095189 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-utilities\") pod \"05a64682-151b-4988-b15a-b5647ce67e74\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.095248 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-catalog-content\") pod \"05a64682-151b-4988-b15a-b5647ce67e74\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.095330 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc7fs\" (UniqueName: \"kubernetes.io/projected/05a64682-151b-4988-b15a-b5647ce67e74-kube-api-access-kc7fs\") pod \"05a64682-151b-4988-b15a-b5647ce67e74\" (UID: \"05a64682-151b-4988-b15a-b5647ce67e74\") " Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.097693 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-utilities" (OuterVolumeSpecName: "utilities") pod "05a64682-151b-4988-b15a-b5647ce67e74" (UID: "05a64682-151b-4988-b15a-b5647ce67e74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.107699 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05a64682-151b-4988-b15a-b5647ce67e74-kube-api-access-kc7fs" (OuterVolumeSpecName: "kube-api-access-kc7fs") pod "05a64682-151b-4988-b15a-b5647ce67e74" (UID: "05a64682-151b-4988-b15a-b5647ce67e74"). InnerVolumeSpecName "kube-api-access-kc7fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.131942 4914 scope.go:117] "RemoveContainer" containerID="13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.158933 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05a64682-151b-4988-b15a-b5647ce67e74" (UID: "05a64682-151b-4988-b15a-b5647ce67e74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.196387 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc7fs\" (UniqueName: \"kubernetes.io/projected/05a64682-151b-4988-b15a-b5647ce67e74-kube-api-access-kc7fs\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.196421 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.196430 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05a64682-151b-4988-b15a-b5647ce67e74-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.215397 4914 scope.go:117] "RemoveContainer" containerID="06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173" Nov 22 14:30:55 crc kubenswrapper[4914]: E1122 14:30:55.216032 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173\": container with ID starting with 06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173 not found: ID does not exist" containerID="06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.216064 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173"} err="failed to get container status \"06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173\": rpc error: code = NotFound desc = could not find container \"06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173\": container with ID starting with 06748d16fe363e65dc989b252d64757b88b19701edba1b6b6800b4b65e18d173 not found: ID does not exist" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.216085 4914 scope.go:117] "RemoveContainer" containerID="85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4" Nov 22 14:30:55 crc kubenswrapper[4914]: E1122 14:30:55.216381 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4\": container with ID starting with 85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4 not found: ID does not exist" containerID="85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.216414 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4"} err="failed to get container status \"85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4\": rpc error: code = NotFound desc = could not find container \"85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4\": container with ID starting with 85a9123edf2ae359bb414c5c650d38c508a4e85199398f4159e2fac96cced2c4 not found: ID does not exist" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.216431 4914 scope.go:117] "RemoveContainer" containerID="13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2" Nov 22 14:30:55 crc kubenswrapper[4914]: E1122 14:30:55.216632 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2\": container with ID starting with 13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2 not found: ID does not exist" containerID="13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.216657 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2"} err="failed to get container status \"13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2\": rpc error: code = NotFound desc = could not find container \"13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2\": container with ID starting with 13af88bef353e32b18f8be8199f97a2f62e6f190bbc09c657ad1e00f641f90d2 not found: ID does not exist" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.366430 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.366496 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.412001 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mp8qw"] Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.417417 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.421545 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mp8qw"] Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.435148 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.509565 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-utilities\") pod \"26230687-39d3-4cfa-ae80-dd176534e764\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.509625 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9nsh\" (UniqueName: \"kubernetes.io/projected/26230687-39d3-4cfa-ae80-dd176534e764-kube-api-access-w9nsh\") pod \"26230687-39d3-4cfa-ae80-dd176534e764\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.509691 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-catalog-content\") pod \"26230687-39d3-4cfa-ae80-dd176534e764\" (UID: \"26230687-39d3-4cfa-ae80-dd176534e764\") " Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.510600 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-utilities" (OuterVolumeSpecName: "utilities") pod "26230687-39d3-4cfa-ae80-dd176534e764" (UID: "26230687-39d3-4cfa-ae80-dd176534e764"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.513272 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26230687-39d3-4cfa-ae80-dd176534e764-kube-api-access-w9nsh" (OuterVolumeSpecName: "kube-api-access-w9nsh") pod "26230687-39d3-4cfa-ae80-dd176534e764" (UID: "26230687-39d3-4cfa-ae80-dd176534e764"). InnerVolumeSpecName "kube-api-access-w9nsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.560081 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26230687-39d3-4cfa-ae80-dd176534e764" (UID: "26230687-39d3-4cfa-ae80-dd176534e764"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.611556 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.611862 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9nsh\" (UniqueName: \"kubernetes.io/projected/26230687-39d3-4cfa-ae80-dd176534e764-kube-api-access-w9nsh\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.611945 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26230687-39d3-4cfa-ae80-dd176534e764-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:55 crc kubenswrapper[4914]: I1122 14:30:55.646466 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-447c5"] Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.085698 4914 generic.go:334] "Generic (PLEG): container finished" podID="26230687-39d3-4cfa-ae80-dd176534e764" containerID="82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7" exitCode=0 Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.086215 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ld6hj" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.086247 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ld6hj" event={"ID":"26230687-39d3-4cfa-ae80-dd176534e764","Type":"ContainerDied","Data":"82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7"} Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.086331 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ld6hj" event={"ID":"26230687-39d3-4cfa-ae80-dd176534e764","Type":"ContainerDied","Data":"afc514f011cf646d3af3dd456057cab84b26086b163f65a0a565f1fb5385f2b7"} Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.086376 4914 scope.go:117] "RemoveContainer" containerID="82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.106185 4914 scope.go:117] "RemoveContainer" containerID="8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.138010 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ld6hj"] Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.138307 4914 scope.go:117] "RemoveContainer" containerID="5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.142429 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ld6hj"] Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.152348 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.170935 4914 scope.go:117] "RemoveContainer" containerID="82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7" Nov 22 14:30:56 crc kubenswrapper[4914]: E1122 14:30:56.171410 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7\": container with ID starting with 82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7 not found: ID does not exist" containerID="82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.171437 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7"} err="failed to get container status \"82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7\": rpc error: code = NotFound desc = could not find container \"82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7\": container with ID starting with 82d6a3e1ea7f6b72e25670f6163b87a9a728d5c7c029125ed5584b458592ecf7 not found: ID does not exist" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.171458 4914 scope.go:117] "RemoveContainer" containerID="8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde" Nov 22 14:30:56 crc kubenswrapper[4914]: E1122 14:30:56.171936 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde\": container with ID starting with 8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde not found: ID does not exist" containerID="8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.171966 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde"} err="failed to get container status \"8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde\": rpc error: code = NotFound desc = could not find container \"8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde\": container with ID starting with 8bb02e6aaeec5cc78ebe32c816566791ea3af43e868dcd6ea260d66f56831cde not found: ID does not exist" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.171985 4914 scope.go:117] "RemoveContainer" containerID="5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe" Nov 22 14:30:56 crc kubenswrapper[4914]: E1122 14:30:56.172916 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe\": container with ID starting with 5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe not found: ID does not exist" containerID="5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.172978 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe"} err="failed to get container status \"5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe\": rpc error: code = NotFound desc = could not find container \"5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe\": container with ID starting with 5799329bb36d5069418a5f4408bf14009f7ad47b397b0a59c517e12c5585aefe not found: ID does not exist" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.874073 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05a64682-151b-4988-b15a-b5647ce67e74" path="/var/lib/kubelet/pods/05a64682-151b-4988-b15a-b5647ce67e74/volumes" Nov 22 14:30:56 crc kubenswrapper[4914]: I1122 14:30:56.875248 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26230687-39d3-4cfa-ae80-dd176534e764" path="/var/lib/kubelet/pods/26230687-39d3-4cfa-ae80-dd176534e764/volumes" Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.486344 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xk4qr"] Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.487149 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xk4qr" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="registry-server" containerID="cri-o://f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51" gracePeriod=2 Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.846747 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.948759 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-utilities\") pod \"05e12f57-26cf-412d-8a06-0aa535536488\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.948822 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-catalog-content\") pod \"05e12f57-26cf-412d-8a06-0aa535536488\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.948843 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zksbf\" (UniqueName: \"kubernetes.io/projected/05e12f57-26cf-412d-8a06-0aa535536488-kube-api-access-zksbf\") pod \"05e12f57-26cf-412d-8a06-0aa535536488\" (UID: \"05e12f57-26cf-412d-8a06-0aa535536488\") " Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.949540 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-utilities" (OuterVolumeSpecName: "utilities") pod "05e12f57-26cf-412d-8a06-0aa535536488" (UID: "05e12f57-26cf-412d-8a06-0aa535536488"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:58 crc kubenswrapper[4914]: I1122 14:30:58.953654 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e12f57-26cf-412d-8a06-0aa535536488-kube-api-access-zksbf" (OuterVolumeSpecName: "kube-api-access-zksbf") pod "05e12f57-26cf-412d-8a06-0aa535536488" (UID: "05e12f57-26cf-412d-8a06-0aa535536488"). InnerVolumeSpecName "kube-api-access-zksbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.049665 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05e12f57-26cf-412d-8a06-0aa535536488" (UID: "05e12f57-26cf-412d-8a06-0aa535536488"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.050297 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.050313 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05e12f57-26cf-412d-8a06-0aa535536488-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.050323 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zksbf\" (UniqueName: \"kubernetes.io/projected/05e12f57-26cf-412d-8a06-0aa535536488-kube-api-access-zksbf\") on node \"crc\" DevicePath \"\"" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.123408 4914 generic.go:334] "Generic (PLEG): container finished" podID="05e12f57-26cf-412d-8a06-0aa535536488" containerID="f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51" exitCode=0 Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.123460 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerDied","Data":"f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51"} Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.123465 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk4qr" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.123499 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk4qr" event={"ID":"05e12f57-26cf-412d-8a06-0aa535536488","Type":"ContainerDied","Data":"63792ffbf939e4a9000035140107ce8976c3102d9ee6e1b7780b881b09f2d075"} Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.123521 4914 scope.go:117] "RemoveContainer" containerID="f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.149349 4914 scope.go:117] "RemoveContainer" containerID="938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.170243 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xk4qr"] Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.175647 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xk4qr"] Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.176842 4914 scope.go:117] "RemoveContainer" containerID="dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.189830 4914 scope.go:117] "RemoveContainer" containerID="f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51" Nov 22 14:30:59 crc kubenswrapper[4914]: E1122 14:30:59.190269 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51\": container with ID starting with f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51 not found: ID does not exist" containerID="f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.190300 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51"} err="failed to get container status \"f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51\": rpc error: code = NotFound desc = could not find container \"f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51\": container with ID starting with f16c66b3a16eae829daeda643eea7acf3fae268b905994f67bec375fe0ee9d51 not found: ID does not exist" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.190323 4914 scope.go:117] "RemoveContainer" containerID="938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614" Nov 22 14:30:59 crc kubenswrapper[4914]: E1122 14:30:59.190598 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614\": container with ID starting with 938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614 not found: ID does not exist" containerID="938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.190618 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614"} err="failed to get container status \"938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614\": rpc error: code = NotFound desc = could not find container \"938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614\": container with ID starting with 938377ea416ed55048f8425dd91a7edcb91f3f3834a930d2409df085bbcdf614 not found: ID does not exist" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.190630 4914 scope.go:117] "RemoveContainer" containerID="dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c" Nov 22 14:30:59 crc kubenswrapper[4914]: E1122 14:30:59.190972 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c\": container with ID starting with dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c not found: ID does not exist" containerID="dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c" Nov 22 14:30:59 crc kubenswrapper[4914]: I1122 14:30:59.191018 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c"} err="failed to get container status \"dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c\": rpc error: code = NotFound desc = could not find container \"dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c\": container with ID starting with dd0cf106529031c2071518a325c0ba9c5805d9c0be8a5b86ef45efd9b8a7ed7c not found: ID does not exist" Nov 22 14:31:00 crc kubenswrapper[4914]: I1122 14:31:00.877677 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05e12f57-26cf-412d-8a06-0aa535536488" path="/var/lib/kubelet/pods/05e12f57-26cf-412d-8a06-0aa535536488/volumes" Nov 22 14:31:20 crc kubenswrapper[4914]: I1122 14:31:20.676558 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" podUID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" containerName="oauth-openshift" containerID="cri-o://5ddced9750da27bf2c5e12dd8a983fd2097fa5c2d1bbce94fe2eee14d2e3e1a0" gracePeriod=15 Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.253409 4914 generic.go:334] "Generic (PLEG): container finished" podID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" containerID="5ddced9750da27bf2c5e12dd8a983fd2097fa5c2d1bbce94fe2eee14d2e3e1a0" exitCode=0 Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.253457 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" event={"ID":"3afd1e74-f3ea-4c65-b46c-51156ed6ae48","Type":"ContainerDied","Data":"5ddced9750da27bf2c5e12dd8a983fd2097fa5c2d1bbce94fe2eee14d2e3e1a0"} Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.607229 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652298 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-96d6999f9-vr9lx"] Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652580 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652599 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652617 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652629 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652643 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0802ab8f-68e3-41fa-b454-cbbdedd10f29" containerName="collect-profiles" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652655 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="0802ab8f-68e3-41fa-b454-cbbdedd10f29" containerName="collect-profiles" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652670 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652680 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652693 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652703 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652723 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652733 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652749 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652760 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652780 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652791 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652807 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652818 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="extract-utilities" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652833 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652845 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652861 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" containerName="oauth-openshift" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652872 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" containerName="oauth-openshift" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652911 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652922 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652937 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652948 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="extract-content" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652961 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652970 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: E1122 14:31:21.652988 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53837311-c646-4085-b1ea-1a24fb53f621" containerName="pruner" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.652999 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="53837311-c646-4085-b1ea-1a24fb53f621" containerName="pruner" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656023 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="53837311-c646-4085-b1ea-1a24fb53f621" containerName="pruner" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656088 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="05a64682-151b-4988-b15a-b5647ce67e74" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656107 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="0802ab8f-68e3-41fa-b454-cbbdedd10f29" containerName="collect-profiles" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656129 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e12f57-26cf-412d-8a06-0aa535536488" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656144 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="26230687-39d3-4cfa-ae80-dd176534e764" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656169 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" containerName="oauth-openshift" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.656222 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="048fe25e-6b0e-4371-879e-d0b407397aa8" containerName="registry-server" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.657027 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.662321 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-96d6999f9-vr9lx"] Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772228 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-dir\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772308 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-cliconfig\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772332 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-serving-cert\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772344 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772360 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-service-ca\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772458 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-session\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772537 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-trusted-ca-bundle\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772600 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-provider-selection\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772655 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-idp-0-file-data\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772723 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-error\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.772834 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-ocp-branding-template\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.773098 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-policies\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.773267 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.773836 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.773861 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774091 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vvfv\" (UniqueName: \"kubernetes.io/projected/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-kube-api-access-7vvfv\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774182 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-router-certs\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774253 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-login\") pod \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\" (UID: \"3afd1e74-f3ea-4c65-b46c-51156ed6ae48\") " Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774296 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774564 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-cliconfig\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774650 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-audit-policies\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774761 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-login\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774915 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.774989 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-service-ca\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775020 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5236d3c1-aa22-4697-9e94-58883bd7a298-audit-dir\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775045 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-serving-cert\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775074 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-session\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775101 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-router-certs\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775127 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpbf2\" (UniqueName: \"kubernetes.io/projected/5236d3c1-aa22-4697-9e94-58883bd7a298-kube-api-access-hpbf2\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775191 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775278 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775340 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-error\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775369 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775415 4914 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775433 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775447 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775460 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.775474 4914 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.778364 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.778696 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.779849 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.780641 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.780642 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-kube-api-access-7vvfv" (OuterVolumeSpecName: "kube-api-access-7vvfv") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "kube-api-access-7vvfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.780940 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.782036 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.787133 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.787495 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3afd1e74-f3ea-4c65-b46c-51156ed6ae48" (UID: "3afd1e74-f3ea-4c65-b46c-51156ed6ae48"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876261 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876361 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876459 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-error\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876510 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876546 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-cliconfig\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876595 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-audit-policies\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876654 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-login\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876739 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876774 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-service-ca\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876804 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5236d3c1-aa22-4697-9e94-58883bd7a298-audit-dir\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876836 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-serving-cert\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876866 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-session\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876933 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-router-certs\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.876970 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpbf2\" (UniqueName: \"kubernetes.io/projected/5236d3c1-aa22-4697-9e94-58883bd7a298-kube-api-access-hpbf2\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.877058 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5236d3c1-aa22-4697-9e94-58883bd7a298-audit-dir\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.877287 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879443 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-audit-policies\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879627 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-cliconfig\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879651 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879696 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879732 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879769 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vvfv\" (UniqueName: \"kubernetes.io/projected/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-kube-api-access-7vvfv\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879798 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879827 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879861 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879864 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-service-ca\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.879930 4914 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3afd1e74-f3ea-4c65-b46c-51156ed6ae48-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.880107 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.880723 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.880993 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.882984 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-error\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.883082 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.883374 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-serving-cert\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.883492 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-router-certs\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.885999 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-system-session\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.886574 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5236d3c1-aa22-4697-9e94-58883bd7a298-v4-0-config-user-template-login\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.894722 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpbf2\" (UniqueName: \"kubernetes.io/projected/5236d3c1-aa22-4697-9e94-58883bd7a298-kube-api-access-hpbf2\") pod \"oauth-openshift-96d6999f9-vr9lx\" (UID: \"5236d3c1-aa22-4697-9e94-58883bd7a298\") " pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:21 crc kubenswrapper[4914]: I1122 14:31:21.979190 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.212835 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-96d6999f9-vr9lx"] Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.259811 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" event={"ID":"5236d3c1-aa22-4697-9e94-58883bd7a298","Type":"ContainerStarted","Data":"33d6854f0c83f4b2b821093ea4cbdb15aa8648aec45610c574ddec5658695b46"} Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.261228 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" event={"ID":"3afd1e74-f3ea-4c65-b46c-51156ed6ae48","Type":"ContainerDied","Data":"fadc0c987b740a4025e54b9c01aa0a2eeac6adf5ea8be506dacd792bace5f5de"} Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.261272 4914 scope.go:117] "RemoveContainer" containerID="5ddced9750da27bf2c5e12dd8a983fd2097fa5c2d1bbce94fe2eee14d2e3e1a0" Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.261401 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-447c5" Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.299188 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-447c5"] Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.301405 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-447c5"] Nov 22 14:31:22 crc kubenswrapper[4914]: I1122 14:31:22.884814 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3afd1e74-f3ea-4c65-b46c-51156ed6ae48" path="/var/lib/kubelet/pods/3afd1e74-f3ea-4c65-b46c-51156ed6ae48/volumes" Nov 22 14:31:23 crc kubenswrapper[4914]: I1122 14:31:23.266620 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" event={"ID":"5236d3c1-aa22-4697-9e94-58883bd7a298","Type":"ContainerStarted","Data":"8785aeada6c419dcad0de448e4ea3b08167c49f519d566610bf70cfb1f2a8792"} Nov 22 14:31:23 crc kubenswrapper[4914]: I1122 14:31:23.267019 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:23 crc kubenswrapper[4914]: I1122 14:31:23.272531 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" Nov 22 14:31:23 crc kubenswrapper[4914]: I1122 14:31:23.287790 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-96d6999f9-vr9lx" podStartSLOduration=28.287774691 podStartE2EDuration="28.287774691s" podCreationTimestamp="2025-11-22 14:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:31:23.284728472 +0000 UTC m=+249.093298143" watchObservedRunningTime="2025-11-22 14:31:23.287774691 +0000 UTC m=+249.096344352" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.014987 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m54pn"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.015864 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m54pn" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="registry-server" containerID="cri-o://443a98405c3fdea9a0192100b4ed00005cb292baa11971410303508953c88e8f" gracePeriod=30 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.027612 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvnps"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.028210 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xvnps" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="registry-server" containerID="cri-o://cf86376eeb5c3a7e88bfa303595d94c75a66a6c1796ea31feb52793efc756995" gracePeriod=30 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.055084 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kdkmv"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.064770 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" containerID="cri-o://ee809704a7a393b86e908be334d87adce0e5a56685efccb2b18d5e45990c8558" gracePeriod=30 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.068998 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hvrzj"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.069765 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.071624 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hvrzj"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.076186 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qb4hc"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.076411 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qb4hc" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="registry-server" containerID="cri-o://fca1a112a12969e6881ba69fc04107601a4fa37a917944d1cd95c19b0b3a80f7" gracePeriod=30 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.078719 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-99q7s"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.078956 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-99q7s" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="registry-server" containerID="cri-o://f4285cff2fb23b58c1f5c16573f8a8a6b38e183919c8e79edd6a755b681b943f" gracePeriod=30 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.202749 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/096f3cee-1a5e-4891-8b45-f1c00531807b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.202850 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/096f3cee-1a5e-4891-8b45-f1c00531807b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.202904 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6kgd\" (UniqueName: \"kubernetes.io/projected/096f3cee-1a5e-4891-8b45-f1c00531807b-kube-api-access-j6kgd\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.304783 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/096f3cee-1a5e-4891-8b45-f1c00531807b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.304951 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/096f3cee-1a5e-4891-8b45-f1c00531807b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.304996 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6kgd\" (UniqueName: \"kubernetes.io/projected/096f3cee-1a5e-4891-8b45-f1c00531807b-kube-api-access-j6kgd\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.306414 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/096f3cee-1a5e-4891-8b45-f1c00531807b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.314845 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/096f3cee-1a5e-4891-8b45-f1c00531807b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.327341 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6kgd\" (UniqueName: \"kubernetes.io/projected/096f3cee-1a5e-4891-8b45-f1c00531807b-kube-api-access-j6kgd\") pod \"marketplace-operator-79b997595-hvrzj\" (UID: \"096f3cee-1a5e-4891-8b45-f1c00531807b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.341930 4914 generic.go:334] "Generic (PLEG): container finished" podID="87dd815e-51f0-4214-98f9-2d797a57931a" containerID="f4285cff2fb23b58c1f5c16573f8a8a6b38e183919c8e79edd6a755b681b943f" exitCode=0 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.342002 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerDied","Data":"f4285cff2fb23b58c1f5c16573f8a8a6b38e183919c8e79edd6a755b681b943f"} Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.343796 4914 generic.go:334] "Generic (PLEG): container finished" podID="8275d661-2656-4e16-bff2-40633bdde977" containerID="cf86376eeb5c3a7e88bfa303595d94c75a66a6c1796ea31feb52793efc756995" exitCode=0 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.343846 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnps" event={"ID":"8275d661-2656-4e16-bff2-40633bdde977","Type":"ContainerDied","Data":"cf86376eeb5c3a7e88bfa303595d94c75a66a6c1796ea31feb52793efc756995"} Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.345184 4914 generic.go:334] "Generic (PLEG): container finished" podID="406cdf66-b685-46a9-b879-89bad870a774" containerID="ee809704a7a393b86e908be334d87adce0e5a56685efccb2b18d5e45990c8558" exitCode=0 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.345260 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" event={"ID":"406cdf66-b685-46a9-b879-89bad870a774","Type":"ContainerDied","Data":"ee809704a7a393b86e908be334d87adce0e5a56685efccb2b18d5e45990c8558"} Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.346974 4914 generic.go:334] "Generic (PLEG): container finished" podID="35766d87-0764-410c-ba6c-e0afc70a0379" containerID="443a98405c3fdea9a0192100b4ed00005cb292baa11971410303508953c88e8f" exitCode=0 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.347019 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerDied","Data":"443a98405c3fdea9a0192100b4ed00005cb292baa11971410303508953c88e8f"} Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.348486 4914 generic.go:334] "Generic (PLEG): container finished" podID="f4dd5c74-ad7f-4740-91c1-604264249996" containerID="fca1a112a12969e6881ba69fc04107601a4fa37a917944d1cd95c19b0b3a80f7" exitCode=0 Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.348517 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerDied","Data":"fca1a112a12969e6881ba69fc04107601a4fa37a917944d1cd95c19b0b3a80f7"} Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.430117 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.534987 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.610562 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-utilities\") pod \"35766d87-0764-410c-ba6c-e0afc70a0379\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.610663 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-catalog-content\") pod \"35766d87-0764-410c-ba6c-e0afc70a0379\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.610691 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crndh\" (UniqueName: \"kubernetes.io/projected/35766d87-0764-410c-ba6c-e0afc70a0379-kube-api-access-crndh\") pod \"35766d87-0764-410c-ba6c-e0afc70a0379\" (UID: \"35766d87-0764-410c-ba6c-e0afc70a0379\") " Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.611458 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-utilities" (OuterVolumeSpecName: "utilities") pod "35766d87-0764-410c-ba6c-e0afc70a0379" (UID: "35766d87-0764-410c-ba6c-e0afc70a0379"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.615962 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35766d87-0764-410c-ba6c-e0afc70a0379-kube-api-access-crndh" (OuterVolumeSpecName: "kube-api-access-crndh") pod "35766d87-0764-410c-ba6c-e0afc70a0379" (UID: "35766d87-0764-410c-ba6c-e0afc70a0379"). InnerVolumeSpecName "kube-api-access-crndh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.658088 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35766d87-0764-410c-ba6c-e0afc70a0379" (UID: "35766d87-0764-410c-ba6c-e0afc70a0379"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.712144 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.712178 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crndh\" (UniqueName: \"kubernetes.io/projected/35766d87-0764-410c-ba6c-e0afc70a0379-kube-api-access-crndh\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.712191 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35766d87-0764-410c-ba6c-e0afc70a0379-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.753851 4914 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kdkmv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.753928 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.836264 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.839182 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hvrzj"] Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.914000 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-utilities\") pod \"8275d661-2656-4e16-bff2-40633bdde977\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.914635 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-catalog-content\") pod \"8275d661-2656-4e16-bff2-40633bdde977\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.914721 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z7nd\" (UniqueName: \"kubernetes.io/projected/8275d661-2656-4e16-bff2-40633bdde977-kube-api-access-6z7nd\") pod \"8275d661-2656-4e16-bff2-40633bdde977\" (UID: \"8275d661-2656-4e16-bff2-40633bdde977\") " Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.914832 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-utilities" (OuterVolumeSpecName: "utilities") pod "8275d661-2656-4e16-bff2-40633bdde977" (UID: "8275d661-2656-4e16-bff2-40633bdde977"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.915047 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.932309 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8275d661-2656-4e16-bff2-40633bdde977-kube-api-access-6z7nd" (OuterVolumeSpecName: "kube-api-access-6z7nd") pod "8275d661-2656-4e16-bff2-40633bdde977" (UID: "8275d661-2656-4e16-bff2-40633bdde977"). InnerVolumeSpecName "kube-api-access-6z7nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:31:36 crc kubenswrapper[4914]: I1122 14:31:36.972461 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.000604 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.004376 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.009328 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8275d661-2656-4e16-bff2-40633bdde977" (UID: "8275d661-2656-4e16-bff2-40633bdde977"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.015674 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/406cdf66-b685-46a9-b879-89bad870a774-marketplace-trusted-ca\") pod \"406cdf66-b685-46a9-b879-89bad870a774\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.016731 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmrg8\" (UniqueName: \"kubernetes.io/projected/406cdf66-b685-46a9-b879-89bad870a774-kube-api-access-mmrg8\") pod \"406cdf66-b685-46a9-b879-89bad870a774\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.016919 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/406cdf66-b685-46a9-b879-89bad870a774-marketplace-operator-metrics\") pod \"406cdf66-b685-46a9-b879-89bad870a774\" (UID: \"406cdf66-b685-46a9-b879-89bad870a774\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.017247 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8275d661-2656-4e16-bff2-40633bdde977-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.017318 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z7nd\" (UniqueName: \"kubernetes.io/projected/8275d661-2656-4e16-bff2-40633bdde977-kube-api-access-6z7nd\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.016430 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406cdf66-b685-46a9-b879-89bad870a774-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "406cdf66-b685-46a9-b879-89bad870a774" (UID: "406cdf66-b685-46a9-b879-89bad870a774"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.020732 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406cdf66-b685-46a9-b879-89bad870a774-kube-api-access-mmrg8" (OuterVolumeSpecName: "kube-api-access-mmrg8") pod "406cdf66-b685-46a9-b879-89bad870a774" (UID: "406cdf66-b685-46a9-b879-89bad870a774"). InnerVolumeSpecName "kube-api-access-mmrg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.022632 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406cdf66-b685-46a9-b879-89bad870a774-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "406cdf66-b685-46a9-b879-89bad870a774" (UID: "406cdf66-b685-46a9-b879-89bad870a774"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118500 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-catalog-content\") pod \"f4dd5c74-ad7f-4740-91c1-604264249996\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118558 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-utilities\") pod \"f4dd5c74-ad7f-4740-91c1-604264249996\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118579 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-catalog-content\") pod \"87dd815e-51f0-4214-98f9-2d797a57931a\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118629 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85f24\" (UniqueName: \"kubernetes.io/projected/87dd815e-51f0-4214-98f9-2d797a57931a-kube-api-access-85f24\") pod \"87dd815e-51f0-4214-98f9-2d797a57931a\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118675 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2jlq\" (UniqueName: \"kubernetes.io/projected/f4dd5c74-ad7f-4740-91c1-604264249996-kube-api-access-p2jlq\") pod \"f4dd5c74-ad7f-4740-91c1-604264249996\" (UID: \"f4dd5c74-ad7f-4740-91c1-604264249996\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118696 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-utilities\") pod \"87dd815e-51f0-4214-98f9-2d797a57931a\" (UID: \"87dd815e-51f0-4214-98f9-2d797a57931a\") " Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118867 4914 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/406cdf66-b685-46a9-b879-89bad870a774-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118896 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmrg8\" (UniqueName: \"kubernetes.io/projected/406cdf66-b685-46a9-b879-89bad870a774-kube-api-access-mmrg8\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.118907 4914 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/406cdf66-b685-46a9-b879-89bad870a774-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.119767 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-utilities" (OuterVolumeSpecName: "utilities") pod "f4dd5c74-ad7f-4740-91c1-604264249996" (UID: "f4dd5c74-ad7f-4740-91c1-604264249996"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.120146 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-utilities" (OuterVolumeSpecName: "utilities") pod "87dd815e-51f0-4214-98f9-2d797a57931a" (UID: "87dd815e-51f0-4214-98f9-2d797a57931a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.121747 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4dd5c74-ad7f-4740-91c1-604264249996-kube-api-access-p2jlq" (OuterVolumeSpecName: "kube-api-access-p2jlq") pod "f4dd5c74-ad7f-4740-91c1-604264249996" (UID: "f4dd5c74-ad7f-4740-91c1-604264249996"). InnerVolumeSpecName "kube-api-access-p2jlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.122121 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87dd815e-51f0-4214-98f9-2d797a57931a-kube-api-access-85f24" (OuterVolumeSpecName: "kube-api-access-85f24") pod "87dd815e-51f0-4214-98f9-2d797a57931a" (UID: "87dd815e-51f0-4214-98f9-2d797a57931a"). InnerVolumeSpecName "kube-api-access-85f24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.136482 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4dd5c74-ad7f-4740-91c1-604264249996" (UID: "f4dd5c74-ad7f-4740-91c1-604264249996"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.213125 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87dd815e-51f0-4214-98f9-2d797a57931a" (UID: "87dd815e-51f0-4214-98f9-2d797a57931a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.219498 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.219527 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4dd5c74-ad7f-4740-91c1-604264249996-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.219536 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.219546 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85f24\" (UniqueName: \"kubernetes.io/projected/87dd815e-51f0-4214-98f9-2d797a57931a-kube-api-access-85f24\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.219556 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2jlq\" (UniqueName: \"kubernetes.io/projected/f4dd5c74-ad7f-4740-91c1-604264249996-kube-api-access-p2jlq\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.219564 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87dd815e-51f0-4214-98f9-2d797a57931a-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.354501 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" event={"ID":"406cdf66-b685-46a9-b879-89bad870a774","Type":"ContainerDied","Data":"0576c01e1d290f31cd3ec73b6eb15ba60075089c6b9cb92268fe428c9184ed60"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.354795 4914 scope.go:117] "RemoveContainer" containerID="ee809704a7a393b86e908be334d87adce0e5a56685efccb2b18d5e45990c8558" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.354541 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kdkmv" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.356740 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnps" event={"ID":"8275d661-2656-4e16-bff2-40633bdde977","Type":"ContainerDied","Data":"b17c557d9d658e3e8948dd2ac802c4884d2dfc2ff5d6e854f0ff6479fd8182d1"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.356779 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvnps" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.359424 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m54pn" event={"ID":"35766d87-0764-410c-ba6c-e0afc70a0379","Type":"ContainerDied","Data":"1e5fa40e9a58d8765ef22407255be07ac849a21790ae5a5c0f74b3f149c48122"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.359463 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m54pn" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.362359 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qb4hc" event={"ID":"f4dd5c74-ad7f-4740-91c1-604264249996","Type":"ContainerDied","Data":"135517cbc00e1a29ff389b5b88fdbf21b461a57c570ad62736424a9d8b3a8de1"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.362480 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qb4hc" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.365682 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99q7s" event={"ID":"87dd815e-51f0-4214-98f9-2d797a57931a","Type":"ContainerDied","Data":"b249cba2937a88830c07027e44ea4f32eef0c6bf3bbd304df132012f632d1024"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.365751 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99q7s" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.370567 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" event={"ID":"096f3cee-1a5e-4891-8b45-f1c00531807b","Type":"ContainerStarted","Data":"07122774ff811caf60bffcf6235a006a3c2b9444085c7186056b025c41b06f8e"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.370621 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" event={"ID":"096f3cee-1a5e-4891-8b45-f1c00531807b","Type":"ContainerStarted","Data":"1061b26fb2fa2a33b8568e3f41cea719a8f1a0b6517bf5889c07f9838e75b7f5"} Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.371623 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.375510 4914 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hvrzj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.375559 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" podUID="096f3cee-1a5e-4891-8b45-f1c00531807b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.375810 4914 scope.go:117] "RemoveContainer" containerID="cf86376eeb5c3a7e88bfa303595d94c75a66a6c1796ea31feb52793efc756995" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.391375 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m54pn"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.396294 4914 scope.go:117] "RemoveContainer" containerID="5366e763a2591e5a36195e4674dd314f69a78215fb134a19545a0375700c3778" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.396846 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m54pn"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.409699 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvnps"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.412740 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xvnps"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.425511 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" podStartSLOduration=1.425497357 podStartE2EDuration="1.425497357s" podCreationTimestamp="2025-11-22 14:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:31:37.425085336 +0000 UTC m=+263.233654997" watchObservedRunningTime="2025-11-22 14:31:37.425497357 +0000 UTC m=+263.234067018" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.433799 4914 scope.go:117] "RemoveContainer" containerID="15c2d4ad9db2b3f91d1322177f4a741389dcfce1ef144eea444628173edb9ba8" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.445362 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kdkmv"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.467889 4914 scope.go:117] "RemoveContainer" containerID="443a98405c3fdea9a0192100b4ed00005cb292baa11971410303508953c88e8f" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.468045 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kdkmv"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.468076 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qb4hc"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.469771 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qb4hc"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.481342 4914 scope.go:117] "RemoveContainer" containerID="08c3c465ed9e6e79713084ebbfd77f13ad671d644873efbece6e91a602657982" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.489067 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-99q7s"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.492344 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-99q7s"] Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.499199 4914 scope.go:117] "RemoveContainer" containerID="86136a6f8d1ab7b05945b469084ef591c39cea00f082e37a7eb58ec6a2230679" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.525532 4914 scope.go:117] "RemoveContainer" containerID="fca1a112a12969e6881ba69fc04107601a4fa37a917944d1cd95c19b0b3a80f7" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.540519 4914 scope.go:117] "RemoveContainer" containerID="8ee2c0dce00bba6aeb82e1470ef8540119ddd09466828bcd98ef81d65fcd23f9" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.559634 4914 scope.go:117] "RemoveContainer" containerID="b273a102f0c2c8cda400566d38bbf3f516fe132d145ee0cc4cb1bead35e93df7" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.588073 4914 scope.go:117] "RemoveContainer" containerID="f4285cff2fb23b58c1f5c16573f8a8a6b38e183919c8e79edd6a755b681b943f" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.605434 4914 scope.go:117] "RemoveContainer" containerID="7cdb66731fcb66fb1d2f7e1dfc3483f91cde0d2405fcece30f028c9dfe8aa233" Nov 22 14:31:37 crc kubenswrapper[4914]: I1122 14:31:37.620077 4914 scope.go:117] "RemoveContainer" containerID="d83ff8dc643fc00e70b856970ddbc8b13fc10a0d32871e2c88d452d05d46fa02" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232616 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvnv6"] Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232897 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232909 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232921 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232927 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232936 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232942 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232949 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232955 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232963 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232969 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232982 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.232989 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.232997 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233003 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.233018 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233025 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.233033 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233040 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="extract-utilities" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.233049 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233057 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.233066 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233072 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.233082 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233087 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="extract-content" Nov 22 14:31:38 crc kubenswrapper[4914]: E1122 14:31:38.233095 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233100 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233183 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="406cdf66-b685-46a9-b879-89bad870a774" containerName="marketplace-operator" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233192 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="8275d661-2656-4e16-bff2-40633bdde977" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233205 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233212 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.233221 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" containerName="registry-server" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.234044 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.236505 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.247587 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvnv6"] Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.333955 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e533fa-1996-4a43-949d-f5a1ea000dc7-catalog-content\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.334027 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e533fa-1996-4a43-949d-f5a1ea000dc7-utilities\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.334081 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjtkm\" (UniqueName: \"kubernetes.io/projected/73e533fa-1996-4a43-949d-f5a1ea000dc7-kube-api-access-fjtkm\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.382483 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hvrzj" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.425050 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pgm5g"] Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.426710 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.429287 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.434958 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e533fa-1996-4a43-949d-f5a1ea000dc7-catalog-content\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.435020 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e533fa-1996-4a43-949d-f5a1ea000dc7-utilities\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.435084 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjtkm\" (UniqueName: \"kubernetes.io/projected/73e533fa-1996-4a43-949d-f5a1ea000dc7-kube-api-access-fjtkm\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.435674 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e533fa-1996-4a43-949d-f5a1ea000dc7-catalog-content\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.436192 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgm5g"] Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.436570 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e533fa-1996-4a43-949d-f5a1ea000dc7-utilities\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.463862 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjtkm\" (UniqueName: \"kubernetes.io/projected/73e533fa-1996-4a43-949d-f5a1ea000dc7-kube-api-access-fjtkm\") pod \"community-operators-xvnv6\" (UID: \"73e533fa-1996-4a43-949d-f5a1ea000dc7\") " pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.536501 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5gvh\" (UniqueName: \"kubernetes.io/projected/24938bf2-538b-4d51-9ac5-db5947989fa2-kube-api-access-r5gvh\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.536560 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24938bf2-538b-4d51-9ac5-db5947989fa2-utilities\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.536593 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24938bf2-538b-4d51-9ac5-db5947989fa2-catalog-content\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.566363 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.638584 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5gvh\" (UniqueName: \"kubernetes.io/projected/24938bf2-538b-4d51-9ac5-db5947989fa2-kube-api-access-r5gvh\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.638633 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24938bf2-538b-4d51-9ac5-db5947989fa2-utilities\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.638665 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24938bf2-538b-4d51-9ac5-db5947989fa2-catalog-content\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.639242 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24938bf2-538b-4d51-9ac5-db5947989fa2-catalog-content\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.639701 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24938bf2-538b-4d51-9ac5-db5947989fa2-utilities\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.663571 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5gvh\" (UniqueName: \"kubernetes.io/projected/24938bf2-538b-4d51-9ac5-db5947989fa2-kube-api-access-r5gvh\") pod \"redhat-marketplace-pgm5g\" (UID: \"24938bf2-538b-4d51-9ac5-db5947989fa2\") " pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.733327 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvnv6"] Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.750678 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.875630 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35766d87-0764-410c-ba6c-e0afc70a0379" path="/var/lib/kubelet/pods/35766d87-0764-410c-ba6c-e0afc70a0379/volumes" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.876546 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406cdf66-b685-46a9-b879-89bad870a774" path="/var/lib/kubelet/pods/406cdf66-b685-46a9-b879-89bad870a774/volumes" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.876990 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8275d661-2656-4e16-bff2-40633bdde977" path="/var/lib/kubelet/pods/8275d661-2656-4e16-bff2-40633bdde977/volumes" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.878045 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87dd815e-51f0-4214-98f9-2d797a57931a" path="/var/lib/kubelet/pods/87dd815e-51f0-4214-98f9-2d797a57931a/volumes" Nov 22 14:31:38 crc kubenswrapper[4914]: I1122 14:31:38.879001 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4dd5c74-ad7f-4740-91c1-604264249996" path="/var/lib/kubelet/pods/f4dd5c74-ad7f-4740-91c1-604264249996/volumes" Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.115816 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgm5g"] Nov 22 14:31:39 crc kubenswrapper[4914]: W1122 14:31:39.127038 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24938bf2_538b_4d51_9ac5_db5947989fa2.slice/crio-e4a7856c131f17d27410f57a492e03498aff7c6d3514ab46ba9c4bd9a0db21e5 WatchSource:0}: Error finding container e4a7856c131f17d27410f57a492e03498aff7c6d3514ab46ba9c4bd9a0db21e5: Status 404 returned error can't find the container with id e4a7856c131f17d27410f57a492e03498aff7c6d3514ab46ba9c4bd9a0db21e5 Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.383494 4914 generic.go:334] "Generic (PLEG): container finished" podID="73e533fa-1996-4a43-949d-f5a1ea000dc7" containerID="c54b682899ef5f7c27b9e6eb69d5c8ad254346d684facb759fc984b339e4fb19" exitCode=0 Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.383572 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnv6" event={"ID":"73e533fa-1996-4a43-949d-f5a1ea000dc7","Type":"ContainerDied","Data":"c54b682899ef5f7c27b9e6eb69d5c8ad254346d684facb759fc984b339e4fb19"} Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.383614 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnv6" event={"ID":"73e533fa-1996-4a43-949d-f5a1ea000dc7","Type":"ContainerStarted","Data":"3c232e4e834e26f892e87e0ce8249c21f7433515b0cfcc31251542a2ce4871bb"} Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.386413 4914 generic.go:334] "Generic (PLEG): container finished" podID="24938bf2-538b-4d51-9ac5-db5947989fa2" containerID="4d107b58a8e5831f6475c8248c74bb2d9a29c655d60a4e537fc0338032cf3db3" exitCode=0 Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.386964 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgm5g" event={"ID":"24938bf2-538b-4d51-9ac5-db5947989fa2","Type":"ContainerDied","Data":"4d107b58a8e5831f6475c8248c74bb2d9a29c655d60a4e537fc0338032cf3db3"} Nov 22 14:31:39 crc kubenswrapper[4914]: I1122 14:31:39.387066 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgm5g" event={"ID":"24938bf2-538b-4d51-9ac5-db5947989fa2","Type":"ContainerStarted","Data":"e4a7856c131f17d27410f57a492e03498aff7c6d3514ab46ba9c4bd9a0db21e5"} Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.399917 4914 generic.go:334] "Generic (PLEG): container finished" podID="73e533fa-1996-4a43-949d-f5a1ea000dc7" containerID="2948add4acdfea52ff82fb1af2e0b4d99fdf868c80d28f199617e4921ac03ce7" exitCode=0 Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.400102 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnv6" event={"ID":"73e533fa-1996-4a43-949d-f5a1ea000dc7","Type":"ContainerDied","Data":"2948add4acdfea52ff82fb1af2e0b4d99fdf868c80d28f199617e4921ac03ce7"} Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.402918 4914 generic.go:334] "Generic (PLEG): container finished" podID="24938bf2-538b-4d51-9ac5-db5947989fa2" containerID="bf8675d1787dd6c90739985e0280b7dcfc9b7bef3013df8882efbfc1c70a205c" exitCode=0 Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.403346 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgm5g" event={"ID":"24938bf2-538b-4d51-9ac5-db5947989fa2","Type":"ContainerDied","Data":"bf8675d1787dd6c90739985e0280b7dcfc9b7bef3013df8882efbfc1c70a205c"} Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.626252 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lxdjp"] Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.627204 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.635218 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.642542 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxdjp"] Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.663101 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8p2m\" (UniqueName: \"kubernetes.io/projected/a26ac6be-fa89-4b85-89b9-e24dc1065498-kube-api-access-x8p2m\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.663171 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26ac6be-fa89-4b85-89b9-e24dc1065498-utilities\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.663193 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26ac6be-fa89-4b85-89b9-e24dc1065498-catalog-content\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.764518 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8p2m\" (UniqueName: \"kubernetes.io/projected/a26ac6be-fa89-4b85-89b9-e24dc1065498-kube-api-access-x8p2m\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.765057 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26ac6be-fa89-4b85-89b9-e24dc1065498-utilities\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.765083 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26ac6be-fa89-4b85-89b9-e24dc1065498-catalog-content\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.765501 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26ac6be-fa89-4b85-89b9-e24dc1065498-utilities\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.765784 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26ac6be-fa89-4b85-89b9-e24dc1065498-catalog-content\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.796638 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8p2m\" (UniqueName: \"kubernetes.io/projected/a26ac6be-fa89-4b85-89b9-e24dc1065498-kube-api-access-x8p2m\") pod \"certified-operators-lxdjp\" (UID: \"a26ac6be-fa89-4b85-89b9-e24dc1065498\") " pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.827639 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jh4qr"] Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.831368 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.833510 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.836176 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jh4qr"] Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.866672 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn74q\" (UniqueName: \"kubernetes.io/projected/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-kube-api-access-bn74q\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.866782 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-catalog-content\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.866837 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-utilities\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.945966 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.967563 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn74q\" (UniqueName: \"kubernetes.io/projected/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-kube-api-access-bn74q\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.967866 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-catalog-content\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.967945 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-utilities\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.968639 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-utilities\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.968817 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-catalog-content\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:40 crc kubenswrapper[4914]: I1122 14:31:40.985482 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn74q\" (UniqueName: \"kubernetes.io/projected/94fd4ac6-3c90-45c1-9d44-4f8d6977be8b-kube-api-access-bn74q\") pod \"redhat-operators-jh4qr\" (UID: \"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b\") " pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.165722 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.205270 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxdjp"] Nov 22 14:31:41 crc kubenswrapper[4914]: W1122 14:31:41.222076 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda26ac6be_fa89_4b85_89b9_e24dc1065498.slice/crio-b400a0f6b749a81a352953f8e9f8f112e75dde87838486a4e32ccb41983b9ee7 WatchSource:0}: Error finding container b400a0f6b749a81a352953f8e9f8f112e75dde87838486a4e32ccb41983b9ee7: Status 404 returned error can't find the container with id b400a0f6b749a81a352953f8e9f8f112e75dde87838486a4e32ccb41983b9ee7 Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.364842 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jh4qr"] Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.411859 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgm5g" event={"ID":"24938bf2-538b-4d51-9ac5-db5947989fa2","Type":"ContainerStarted","Data":"66007dcd8487c59f6bcd82a63a9680b445a0676a7b636c367489fa069ea28030"} Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.414492 4914 generic.go:334] "Generic (PLEG): container finished" podID="a26ac6be-fa89-4b85-89b9-e24dc1065498" containerID="6335fbee94906d9e7c5c0b031a86678b8ac93ca0af07ffce8d28c76065449a02" exitCode=0 Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.414572 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxdjp" event={"ID":"a26ac6be-fa89-4b85-89b9-e24dc1065498","Type":"ContainerDied","Data":"6335fbee94906d9e7c5c0b031a86678b8ac93ca0af07ffce8d28c76065449a02"} Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.414606 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxdjp" event={"ID":"a26ac6be-fa89-4b85-89b9-e24dc1065498","Type":"ContainerStarted","Data":"b400a0f6b749a81a352953f8e9f8f112e75dde87838486a4e32ccb41983b9ee7"} Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.421236 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvnv6" event={"ID":"73e533fa-1996-4a43-949d-f5a1ea000dc7","Type":"ContainerStarted","Data":"49ae97ded7d496f95ff7dac365df02db61e11539f45b84d4126b47e8d6f68b21"} Nov 22 14:31:41 crc kubenswrapper[4914]: I1122 14:31:41.435998 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pgm5g" podStartSLOduration=2.02545854 podStartE2EDuration="3.435981511s" podCreationTimestamp="2025-11-22 14:31:38 +0000 UTC" firstStartedPulling="2025-11-22 14:31:39.387941269 +0000 UTC m=+265.196510930" lastFinishedPulling="2025-11-22 14:31:40.79846411 +0000 UTC m=+266.607033901" observedRunningTime="2025-11-22 14:31:41.430177089 +0000 UTC m=+267.238746750" watchObservedRunningTime="2025-11-22 14:31:41.435981511 +0000 UTC m=+267.244551172" Nov 22 14:31:42 crc kubenswrapper[4914]: I1122 14:31:42.427727 4914 generic.go:334] "Generic (PLEG): container finished" podID="a26ac6be-fa89-4b85-89b9-e24dc1065498" containerID="e9bf991f69a7583d0633f0c7454f94a8a4fe58756a3f1ab8d703e1672a54b6d9" exitCode=0 Nov 22 14:31:42 crc kubenswrapper[4914]: I1122 14:31:42.427806 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxdjp" event={"ID":"a26ac6be-fa89-4b85-89b9-e24dc1065498","Type":"ContainerDied","Data":"e9bf991f69a7583d0633f0c7454f94a8a4fe58756a3f1ab8d703e1672a54b6d9"} Nov 22 14:31:42 crc kubenswrapper[4914]: I1122 14:31:42.431249 4914 generic.go:334] "Generic (PLEG): container finished" podID="94fd4ac6-3c90-45c1-9d44-4f8d6977be8b" containerID="fb5938cbc0872f173a6caf39e4747b4d4ad9a00711881b054e41d505564363e8" exitCode=0 Nov 22 14:31:42 crc kubenswrapper[4914]: I1122 14:31:42.431284 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jh4qr" event={"ID":"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b","Type":"ContainerDied","Data":"fb5938cbc0872f173a6caf39e4747b4d4ad9a00711881b054e41d505564363e8"} Nov 22 14:31:42 crc kubenswrapper[4914]: I1122 14:31:42.431310 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jh4qr" event={"ID":"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b","Type":"ContainerStarted","Data":"91d6f9378ebc1e82ad651e1a69af796c3599e2f428d7ccae13a14175e3698226"} Nov 22 14:31:42 crc kubenswrapper[4914]: I1122 14:31:42.446346 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvnv6" podStartSLOduration=2.995387951 podStartE2EDuration="4.446323258s" podCreationTimestamp="2025-11-22 14:31:38 +0000 UTC" firstStartedPulling="2025-11-22 14:31:39.384781732 +0000 UTC m=+265.193351393" lastFinishedPulling="2025-11-22 14:31:40.835717039 +0000 UTC m=+266.644286700" observedRunningTime="2025-11-22 14:31:41.468995697 +0000 UTC m=+267.277565378" watchObservedRunningTime="2025-11-22 14:31:42.446323258 +0000 UTC m=+268.254892929" Nov 22 14:31:43 crc kubenswrapper[4914]: I1122 14:31:43.437665 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxdjp" event={"ID":"a26ac6be-fa89-4b85-89b9-e24dc1065498","Type":"ContainerStarted","Data":"71d9dc7aa9409f4cf01e1f7a4224d0dba52edf64fc07e0f4c25cf2c0649343f1"} Nov 22 14:31:43 crc kubenswrapper[4914]: I1122 14:31:43.439678 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jh4qr" event={"ID":"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b","Type":"ContainerStarted","Data":"f5d55a7a05ec4a7934af434ee40819c0def7d6d7d28a1f7836bbfbe463c40375"} Nov 22 14:31:43 crc kubenswrapper[4914]: I1122 14:31:43.456379 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lxdjp" podStartSLOduration=2.005112091 podStartE2EDuration="3.456360508s" podCreationTimestamp="2025-11-22 14:31:40 +0000 UTC" firstStartedPulling="2025-11-22 14:31:41.41630636 +0000 UTC m=+267.224876021" lastFinishedPulling="2025-11-22 14:31:42.867554777 +0000 UTC m=+268.676124438" observedRunningTime="2025-11-22 14:31:43.454946413 +0000 UTC m=+269.263516094" watchObservedRunningTime="2025-11-22 14:31:43.456360508 +0000 UTC m=+269.264930169" Nov 22 14:31:44 crc kubenswrapper[4914]: I1122 14:31:44.450927 4914 generic.go:334] "Generic (PLEG): container finished" podID="94fd4ac6-3c90-45c1-9d44-4f8d6977be8b" containerID="f5d55a7a05ec4a7934af434ee40819c0def7d6d7d28a1f7836bbfbe463c40375" exitCode=0 Nov 22 14:31:44 crc kubenswrapper[4914]: I1122 14:31:44.451048 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jh4qr" event={"ID":"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b","Type":"ContainerDied","Data":"f5d55a7a05ec4a7934af434ee40819c0def7d6d7d28a1f7836bbfbe463c40375"} Nov 22 14:31:45 crc kubenswrapper[4914]: I1122 14:31:45.460588 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jh4qr" event={"ID":"94fd4ac6-3c90-45c1-9d44-4f8d6977be8b","Type":"ContainerStarted","Data":"882ab386a60409aab556d045a4c7757387d5856f2acbc142b39a6e8dc689ad66"} Nov 22 14:31:45 crc kubenswrapper[4914]: I1122 14:31:45.477210 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jh4qr" podStartSLOduration=3.059266168 podStartE2EDuration="5.477193856s" podCreationTimestamp="2025-11-22 14:31:40 +0000 UTC" firstStartedPulling="2025-11-22 14:31:42.438521897 +0000 UTC m=+268.247091558" lastFinishedPulling="2025-11-22 14:31:44.856449585 +0000 UTC m=+270.665019246" observedRunningTime="2025-11-22 14:31:45.474768057 +0000 UTC m=+271.283337718" watchObservedRunningTime="2025-11-22 14:31:45.477193856 +0000 UTC m=+271.285763507" Nov 22 14:31:48 crc kubenswrapper[4914]: I1122 14:31:48.567069 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:48 crc kubenswrapper[4914]: I1122 14:31:48.567448 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:48 crc kubenswrapper[4914]: I1122 14:31:48.627609 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:48 crc kubenswrapper[4914]: I1122 14:31:48.751776 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:48 crc kubenswrapper[4914]: I1122 14:31:48.751838 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:48 crc kubenswrapper[4914]: I1122 14:31:48.814638 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:49 crc kubenswrapper[4914]: I1122 14:31:49.526222 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pgm5g" Nov 22 14:31:49 crc kubenswrapper[4914]: I1122 14:31:49.545814 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvnv6" Nov 22 14:31:50 crc kubenswrapper[4914]: I1122 14:31:50.946718 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:50 crc kubenswrapper[4914]: I1122 14:31:50.946794 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:50 crc kubenswrapper[4914]: I1122 14:31:50.988982 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:31:51 crc kubenswrapper[4914]: I1122 14:31:51.165980 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:51 crc kubenswrapper[4914]: I1122 14:31:51.166058 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:51 crc kubenswrapper[4914]: I1122 14:31:51.221491 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:51 crc kubenswrapper[4914]: I1122 14:31:51.587072 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jh4qr" Nov 22 14:31:51 crc kubenswrapper[4914]: I1122 14:31:51.594841 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lxdjp" Nov 22 14:32:52 crc kubenswrapper[4914]: I1122 14:32:52.042741 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:32:52 crc kubenswrapper[4914]: I1122 14:32:52.043405 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:33:22 crc kubenswrapper[4914]: I1122 14:33:22.043050 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:33:22 crc kubenswrapper[4914]: I1122 14:33:22.043750 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.042804 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.043449 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.043511 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.044331 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"295e4be82247193043ad744eafabe554329435763a26422ff3518c50082e23e2"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.044424 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://295e4be82247193043ad744eafabe554329435763a26422ff3518c50082e23e2" gracePeriod=600 Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.331945 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="295e4be82247193043ad744eafabe554329435763a26422ff3518c50082e23e2" exitCode=0 Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.332018 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"295e4be82247193043ad744eafabe554329435763a26422ff3518c50082e23e2"} Nov 22 14:33:52 crc kubenswrapper[4914]: I1122 14:33:52.332633 4914 scope.go:117] "RemoveContainer" containerID="44fde3bf4fa83439b4d77ccd66dfbe772980aa25ebf33556cfd711ec19b1cc48" Nov 22 14:33:53 crc kubenswrapper[4914]: I1122 14:33:53.344519 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"8bb85fe0e7b9128d7e9a8dfe78ff109dee5730d63a76ca28de3d0e288cea71db"} Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.855705 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bzdgv"] Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.857228 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.874549 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bzdgv"] Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.938481 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.938545 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af7405d-a6fc-4b93-989c-e09618a60096-trusted-ca\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.938581 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7af7405d-a6fc-4b93-989c-e09618a60096-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.938605 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-registry-tls\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.939024 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-bound-sa-token\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.939101 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rbbv\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-kube-api-access-6rbbv\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.939136 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7af7405d-a6fc-4b93-989c-e09618a60096-registry-certificates\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.939169 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7af7405d-a6fc-4b93-989c-e09618a60096-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:27 crc kubenswrapper[4914]: I1122 14:35:27.961444 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.039839 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-bound-sa-token\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.039905 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rbbv\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-kube-api-access-6rbbv\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.039927 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7af7405d-a6fc-4b93-989c-e09618a60096-registry-certificates\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.039951 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7af7405d-a6fc-4b93-989c-e09618a60096-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.039979 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af7405d-a6fc-4b93-989c-e09618a60096-trusted-ca\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.040001 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7af7405d-a6fc-4b93-989c-e09618a60096-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.040020 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-registry-tls\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.040621 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7af7405d-a6fc-4b93-989c-e09618a60096-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.041589 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7af7405d-a6fc-4b93-989c-e09618a60096-registry-certificates\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.042095 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af7405d-a6fc-4b93-989c-e09618a60096-trusted-ca\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.046132 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7af7405d-a6fc-4b93-989c-e09618a60096-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.046297 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-registry-tls\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.058841 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-bound-sa-token\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.060318 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rbbv\" (UniqueName: \"kubernetes.io/projected/7af7405d-a6fc-4b93-989c-e09618a60096-kube-api-access-6rbbv\") pod \"image-registry-66df7c8f76-bzdgv\" (UID: \"7af7405d-a6fc-4b93-989c-e09618a60096\") " pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.179195 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.611493 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bzdgv"] Nov 22 14:35:28 crc kubenswrapper[4914]: W1122 14:35:28.627831 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7af7405d_a6fc_4b93_989c_e09618a60096.slice/crio-51f0164f2c8666ab299a449d91b1d7d85d22cce38011ddecabc2daf94fc452a2 WatchSource:0}: Error finding container 51f0164f2c8666ab299a449d91b1d7d85d22cce38011ddecabc2daf94fc452a2: Status 404 returned error can't find the container with id 51f0164f2c8666ab299a449d91b1d7d85d22cce38011ddecabc2daf94fc452a2 Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.983686 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" event={"ID":"7af7405d-a6fc-4b93-989c-e09618a60096","Type":"ContainerStarted","Data":"3e737a6efc8048f9e93394f665fdf386807b408d3d17462b0d4cae2248b0f715"} Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.983727 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" event={"ID":"7af7405d-a6fc-4b93-989c-e09618a60096","Type":"ContainerStarted","Data":"51f0164f2c8666ab299a449d91b1d7d85d22cce38011ddecabc2daf94fc452a2"} Nov 22 14:35:28 crc kubenswrapper[4914]: I1122 14:35:28.984624 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:29 crc kubenswrapper[4914]: I1122 14:35:29.010735 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" podStartSLOduration=2.01069935 podStartE2EDuration="2.01069935s" podCreationTimestamp="2025-11-22 14:35:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:35:29.006414786 +0000 UTC m=+494.814984447" watchObservedRunningTime="2025-11-22 14:35:29.01069935 +0000 UTC m=+494.819269011" Nov 22 14:35:48 crc kubenswrapper[4914]: I1122 14:35:48.190449 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bzdgv" Nov 22 14:35:48 crc kubenswrapper[4914]: I1122 14:35:48.260462 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mmm9"] Nov 22 14:35:52 crc kubenswrapper[4914]: I1122 14:35:52.043099 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:35:52 crc kubenswrapper[4914]: I1122 14:35:52.044837 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.300550 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" podUID="8f152ee7-771c-47ae-bc31-6e4148a017ef" containerName="registry" containerID="cri-o://68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca" gracePeriod=30 Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.511385 4914 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-8mmm9 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.16:5000/healthz\": dial tcp 10.217.0.16:5000: connect: connection refused" start-of-body= Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.511475 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" podUID="8f152ee7-771c-47ae-bc31-6e4148a017ef" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.16:5000/healthz\": dial tcp 10.217.0.16:5000: connect: connection refused" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.668808 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844314 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-bound-sa-token\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844362 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f152ee7-771c-47ae-bc31-6e4148a017ef-installation-pull-secrets\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844420 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-certificates\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844452 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-trusted-ca\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844596 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844622 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f152ee7-771c-47ae-bc31-6e4148a017ef-ca-trust-extracted\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844653 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-tls\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.844678 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lqr7\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-kube-api-access-2lqr7\") pod \"8f152ee7-771c-47ae-bc31-6e4148a017ef\" (UID: \"8f152ee7-771c-47ae-bc31-6e4148a017ef\") " Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.845285 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.845409 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.852728 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.852787 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f152ee7-771c-47ae-bc31-6e4148a017ef-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.853093 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.854292 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-kube-api-access-2lqr7" (OuterVolumeSpecName: "kube-api-access-2lqr7") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "kube-api-access-2lqr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.861823 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.870426 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f152ee7-771c-47ae-bc31-6e4148a017ef-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f152ee7-771c-47ae-bc31-6e4148a017ef" (UID: "8f152ee7-771c-47ae-bc31-6e4148a017ef"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946303 4914 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946359 4914 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f152ee7-771c-47ae-bc31-6e4148a017ef-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946381 4914 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946400 4914 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f152ee7-771c-47ae-bc31-6e4148a017ef-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946417 4914 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f152ee7-771c-47ae-bc31-6e4148a017ef-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946435 4914 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:13 crc kubenswrapper[4914]: I1122 14:36:13.946453 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lqr7\" (UniqueName: \"kubernetes.io/projected/8f152ee7-771c-47ae-bc31-6e4148a017ef-kube-api-access-2lqr7\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.282527 4914 generic.go:334] "Generic (PLEG): container finished" podID="8f152ee7-771c-47ae-bc31-6e4148a017ef" containerID="68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca" exitCode=0 Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.282572 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" event={"ID":"8f152ee7-771c-47ae-bc31-6e4148a017ef","Type":"ContainerDied","Data":"68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca"} Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.282602 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" event={"ID":"8f152ee7-771c-47ae-bc31-6e4148a017ef","Type":"ContainerDied","Data":"a80abd9ff2771448bf48274d29532040624f6ad26f78087d5d020db263edca7d"} Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.282619 4914 scope.go:117] "RemoveContainer" containerID="68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca" Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.282642 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mmm9" Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.303330 4914 scope.go:117] "RemoveContainer" containerID="68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca" Nov 22 14:36:14 crc kubenswrapper[4914]: E1122 14:36:14.303837 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca\": container with ID starting with 68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca not found: ID does not exist" containerID="68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca" Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.303910 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca"} err="failed to get container status \"68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca\": rpc error: code = NotFound desc = could not find container \"68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca\": container with ID starting with 68d5734ccc140d1cbf999cfe6874d01e6574c28a3c6c79827ad265edb561a4ca not found: ID does not exist" Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.335443 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mmm9"] Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.340488 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mmm9"] Nov 22 14:36:14 crc kubenswrapper[4914]: I1122 14:36:14.880100 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f152ee7-771c-47ae-bc31-6e4148a017ef" path="/var/lib/kubelet/pods/8f152ee7-771c-47ae-bc31-6e4148a017ef/volumes" Nov 22 14:36:22 crc kubenswrapper[4914]: I1122 14:36:22.042938 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:36:22 crc kubenswrapper[4914]: I1122 14:36:22.043588 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.962820 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fftvp"] Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964047 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-controller" containerID="cri-o://659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" gracePeriod=30 Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964106 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="nbdb" containerID="cri-o://32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" gracePeriod=30 Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964213 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-node" containerID="cri-o://15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" gracePeriod=30 Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964239 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="northd" containerID="cri-o://f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" gracePeriod=30 Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964202 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" gracePeriod=30 Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964275 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-acl-logging" containerID="cri-o://fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" gracePeriod=30 Nov 22 14:36:44 crc kubenswrapper[4914]: I1122 14:36:44.964356 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="sbdb" containerID="cri-o://6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" gracePeriod=30 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.012919 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" containerID="cri-o://a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" gracePeriod=30 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.258418 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/3.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.260820 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovn-acl-logging/0.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.261346 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovn-controller/0.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.261855 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312355 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9v8j5"] Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312596 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-node" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312608 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-node" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312618 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312625 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312633 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312640 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312649 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="northd" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312655 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="northd" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312664 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="sbdb" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312670 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="sbdb" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312678 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312683 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312690 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312696 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312704 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312709 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312717 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kubecfg-setup" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312724 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kubecfg-setup" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312732 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="nbdb" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312737 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="nbdb" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312745 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312750 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312758 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f152ee7-771c-47ae-bc31-6e4148a017ef" containerName="registry" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312763 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f152ee7-771c-47ae-bc31-6e4148a017ef" containerName="registry" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.312771 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-acl-logging" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312776 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-acl-logging" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312855 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-ovn-metrics" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312866 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312886 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="kube-rbac-proxy-node" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312896 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="sbdb" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312902 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312909 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovn-acl-logging" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312916 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312923 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="northd" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312931 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312938 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312944 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f152ee7-771c-47ae-bc31-6e4148a017ef" containerName="registry" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.312952 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="nbdb" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.313038 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.313045 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.313126 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" containerName="ovnkube-controller" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.314496 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390662 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-ovn\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390712 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j94x\" (UniqueName: \"kubernetes.io/projected/41ea1017-213b-4be4-87ae-d68655d58a12-kube-api-access-6j94x\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390747 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-slash\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390766 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-systemd\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390781 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-kubelet\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390799 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-var-lib-openvswitch\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390789 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390823 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-netns\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390844 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390855 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ea1017-213b-4be4-87ae-d68655d58a12-ovn-node-metrics-cert\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390915 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-var-lib-cni-networks-ovn-kubernetes\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390906 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390937 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-bin\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390951 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-node-log\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.390999 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-config\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391021 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-netd\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391042 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-ovn-kubernetes\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391028 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391069 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391098 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391059 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-systemd-units\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391129 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391149 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391154 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-openvswitch\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391154 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391169 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-node-log" (OuterVolumeSpecName: "node-log") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391202 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-log-socket\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391231 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391250 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-script-lib\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391279 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-etc-openvswitch\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391297 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-env-overrides\") pod \"41ea1017-213b-4be4-87ae-d68655d58a12\" (UID: \"41ea1017-213b-4be4-87ae-d68655d58a12\") " Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391335 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-log-socket" (OuterVolumeSpecName: "log-socket") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391414 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391654 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391667 4914 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391687 4914 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391699 4914 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-node-log\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391711 4914 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391722 4914 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391734 4914 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391742 4914 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391750 4914 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-log-socket\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391757 4914 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391765 4914 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391772 4914 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391780 4914 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391788 4914 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391831 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-slash" (OuterVolumeSpecName: "host-slash") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.391941 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.392068 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.396553 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ea1017-213b-4be4-87ae-d68655d58a12-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.396581 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ea1017-213b-4be4-87ae-d68655d58a12-kube-api-access-6j94x" (OuterVolumeSpecName: "kube-api-access-6j94x") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "kube-api-access-6j94x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.403689 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "41ea1017-213b-4be4-87ae-d68655d58a12" (UID: "41ea1017-213b-4be4-87ae-d68655d58a12"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.467332 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovnkube-controller/3.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.469924 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovn-acl-logging/0.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.470487 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fftvp_41ea1017-213b-4be4-87ae-d68655d58a12/ovn-controller/0.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471081 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" exitCode=0 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471124 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" exitCode=0 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471136 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" exitCode=0 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471145 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" exitCode=0 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471154 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" exitCode=0 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471162 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" exitCode=0 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471170 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" exitCode=143 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471181 4914 generic.go:334] "Generic (PLEG): container finished" podID="41ea1017-213b-4be4-87ae-d68655d58a12" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" exitCode=143 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471265 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471309 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471326 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471339 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471359 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471373 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471386 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471400 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471406 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471413 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471420 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471430 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471436 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471443 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471449 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471459 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471469 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471477 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471483 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471490 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471497 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471503 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471510 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471516 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471525 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471537 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471547 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471557 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471565 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471572 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471579 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471588 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471595 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471602 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471609 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471615 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471622 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471631 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" event={"ID":"41ea1017-213b-4be4-87ae-d68655d58a12","Type":"ContainerDied","Data":"e19ab2708505489d2bcef86f714e371f5c5d1f20bada0282c2c41b825e9ae44d"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471646 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471654 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471663 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471670 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471677 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471684 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471691 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471698 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471705 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471712 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.471729 4914 scope.go:117] "RemoveContainer" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.472000 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fftvp" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.475350 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/2.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.476067 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/1.log" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.476120 4914 generic.go:334] "Generic (PLEG): container finished" podID="b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed" containerID="bdb8ada03e70d054a4dfb6fa8af17dd17837b11a178753fb893b8fa6403e4862" exitCode=2 Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.476163 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerDied","Data":"bdb8ada03e70d054a4dfb6fa8af17dd17837b11a178753fb893b8fa6403e4862"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.476198 4914 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879"} Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.476754 4914 scope.go:117] "RemoveContainer" containerID="bdb8ada03e70d054a4dfb6fa8af17dd17837b11a178753fb893b8fa6403e4862" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.477080 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8hpjt_openshift-multus(b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed)\"" pod="openshift-multus/multus-8hpjt" podUID="b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.492852 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-cni-netd\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493128 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovnkube-config\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493173 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-run-ovn-kubernetes\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493259 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovnkube-script-lib\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493281 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-slash\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493298 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-systemd\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493330 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-ovn\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493353 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-kubelet\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493382 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-systemd-units\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493466 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-node-log\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493489 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-log-socket\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493504 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovn-node-metrics-cert\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493522 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-var-lib-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493583 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-cni-bin\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493616 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-run-netns\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493652 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccq2h\" (UniqueName: \"kubernetes.io/projected/c8e91974-4a2a-4b81-ae98-10c3ab849f57-kube-api-access-ccq2h\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493677 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493694 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493713 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-env-overrides\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493805 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-etc-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493871 4914 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493906 4914 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493916 4914 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41ea1017-213b-4be4-87ae-d68655d58a12-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493926 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j94x\" (UniqueName: \"kubernetes.io/projected/41ea1017-213b-4be4-87ae-d68655d58a12-kube-api-access-6j94x\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493935 4914 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-host-slash\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493944 4914 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41ea1017-213b-4be4-87ae-d68655d58a12-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.493954 4914 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41ea1017-213b-4be4-87ae-d68655d58a12-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.510119 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.533229 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fftvp"] Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.537234 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fftvp"] Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.546800 4914 scope.go:117] "RemoveContainer" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.566147 4914 scope.go:117] "RemoveContainer" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.585121 4914 scope.go:117] "RemoveContainer" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594481 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccq2h\" (UniqueName: \"kubernetes.io/projected/c8e91974-4a2a-4b81-ae98-10c3ab849f57-kube-api-access-ccq2h\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594527 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594555 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594570 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-env-overrides\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594603 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-etc-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594631 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-cni-netd\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594651 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovnkube-config\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594671 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-run-ovn-kubernetes\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594690 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovnkube-script-lib\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594709 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-slash\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594724 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-systemd\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594716 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594745 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-ovn\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594761 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-kubelet\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594777 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-systemd-units\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594793 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-node-log\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594815 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-log-socket\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594829 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovn-node-metrics-cert\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594847 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-var-lib-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594862 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-cni-bin\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594892 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-run-netns\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594911 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-slash\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594956 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-run-netns\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594967 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.594990 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-systemd\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595015 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-run-ovn\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595034 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-kubelet\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595053 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-systemd-units\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595071 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-node-log\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595112 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-log-socket\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595230 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-var-lib-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595605 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-env-overrides\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595654 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-run-ovn-kubernetes\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595711 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-cni-bin\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595792 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-etc-openvswitch\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.595824 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8e91974-4a2a-4b81-ae98-10c3ab849f57-host-cni-netd\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.596314 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovnkube-config\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.596486 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovnkube-script-lib\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.598501 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8e91974-4a2a-4b81-ae98-10c3ab849f57-ovn-node-metrics-cert\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.605512 4914 scope.go:117] "RemoveContainer" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.620466 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccq2h\" (UniqueName: \"kubernetes.io/projected/c8e91974-4a2a-4b81-ae98-10c3ab849f57-kube-api-access-ccq2h\") pod \"ovnkube-node-9v8j5\" (UID: \"c8e91974-4a2a-4b81-ae98-10c3ab849f57\") " pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.621512 4914 scope.go:117] "RemoveContainer" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.627135 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.635375 4914 scope.go:117] "RemoveContainer" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.656658 4914 scope.go:117] "RemoveContainer" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" Nov 22 14:36:45 crc kubenswrapper[4914]: W1122 14:36:45.668664 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8e91974_4a2a_4b81_ae98_10c3ab849f57.slice/crio-668c87e31cca07fbe153e80405eea04e48b46f7f2d2a22c2a311902d8690ad0e WatchSource:0}: Error finding container 668c87e31cca07fbe153e80405eea04e48b46f7f2d2a22c2a311902d8690ad0e: Status 404 returned error can't find the container with id 668c87e31cca07fbe153e80405eea04e48b46f7f2d2a22c2a311902d8690ad0e Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.673863 4914 scope.go:117] "RemoveContainer" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.704020 4914 scope.go:117] "RemoveContainer" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.704434 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": container with ID starting with a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5 not found: ID does not exist" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.704468 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} err="failed to get container status \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": rpc error: code = NotFound desc = could not find container \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": container with ID starting with a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.704490 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.705082 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": container with ID starting with 8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9 not found: ID does not exist" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.705167 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} err="failed to get container status \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": rpc error: code = NotFound desc = could not find container \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": container with ID starting with 8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.705217 4914 scope.go:117] "RemoveContainer" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.705648 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": container with ID starting with 6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09 not found: ID does not exist" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.705675 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} err="failed to get container status \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": rpc error: code = NotFound desc = could not find container \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": container with ID starting with 6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.705691 4914 scope.go:117] "RemoveContainer" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.706075 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": container with ID starting with 32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f not found: ID does not exist" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.706120 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} err="failed to get container status \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": rpc error: code = NotFound desc = could not find container \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": container with ID starting with 32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.706149 4914 scope.go:117] "RemoveContainer" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.706691 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": container with ID starting with f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4 not found: ID does not exist" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.706752 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} err="failed to get container status \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": rpc error: code = NotFound desc = could not find container \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": container with ID starting with f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.706807 4914 scope.go:117] "RemoveContainer" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.707305 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": container with ID starting with e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d not found: ID does not exist" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.707329 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} err="failed to get container status \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": rpc error: code = NotFound desc = could not find container \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": container with ID starting with e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.707344 4914 scope.go:117] "RemoveContainer" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.707776 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": container with ID starting with 15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde not found: ID does not exist" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.707836 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} err="failed to get container status \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": rpc error: code = NotFound desc = could not find container \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": container with ID starting with 15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.707867 4914 scope.go:117] "RemoveContainer" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.708232 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": container with ID starting with fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024 not found: ID does not exist" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.708259 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} err="failed to get container status \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": rpc error: code = NotFound desc = could not find container \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": container with ID starting with fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.708273 4914 scope.go:117] "RemoveContainer" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.709247 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": container with ID starting with 659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de not found: ID does not exist" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.709299 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} err="failed to get container status \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": rpc error: code = NotFound desc = could not find container \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": container with ID starting with 659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.709330 4914 scope.go:117] "RemoveContainer" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" Nov 22 14:36:45 crc kubenswrapper[4914]: E1122 14:36:45.709846 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": container with ID starting with 0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe not found: ID does not exist" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.709887 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} err="failed to get container status \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": rpc error: code = NotFound desc = could not find container \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": container with ID starting with 0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.709902 4914 scope.go:117] "RemoveContainer" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.710406 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} err="failed to get container status \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": rpc error: code = NotFound desc = could not find container \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": container with ID starting with a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.710426 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.710785 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} err="failed to get container status \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": rpc error: code = NotFound desc = could not find container \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": container with ID starting with 8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.710823 4914 scope.go:117] "RemoveContainer" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.711238 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} err="failed to get container status \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": rpc error: code = NotFound desc = could not find container \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": container with ID starting with 6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.711299 4914 scope.go:117] "RemoveContainer" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.712219 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} err="failed to get container status \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": rpc error: code = NotFound desc = could not find container \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": container with ID starting with 32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.712260 4914 scope.go:117] "RemoveContainer" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.712584 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} err="failed to get container status \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": rpc error: code = NotFound desc = could not find container \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": container with ID starting with f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.712618 4914 scope.go:117] "RemoveContainer" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.713152 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} err="failed to get container status \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": rpc error: code = NotFound desc = could not find container \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": container with ID starting with e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.713174 4914 scope.go:117] "RemoveContainer" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.713484 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} err="failed to get container status \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": rpc error: code = NotFound desc = could not find container \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": container with ID starting with 15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.713523 4914 scope.go:117] "RemoveContainer" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.713779 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} err="failed to get container status \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": rpc error: code = NotFound desc = could not find container \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": container with ID starting with fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.713798 4914 scope.go:117] "RemoveContainer" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.714142 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} err="failed to get container status \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": rpc error: code = NotFound desc = could not find container \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": container with ID starting with 659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.714161 4914 scope.go:117] "RemoveContainer" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.714772 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} err="failed to get container status \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": rpc error: code = NotFound desc = could not find container \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": container with ID starting with 0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.714822 4914 scope.go:117] "RemoveContainer" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.715332 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} err="failed to get container status \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": rpc error: code = NotFound desc = could not find container \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": container with ID starting with a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.715371 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.715703 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} err="failed to get container status \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": rpc error: code = NotFound desc = could not find container \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": container with ID starting with 8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.715730 4914 scope.go:117] "RemoveContainer" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.716045 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} err="failed to get container status \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": rpc error: code = NotFound desc = could not find container \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": container with ID starting with 6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.716085 4914 scope.go:117] "RemoveContainer" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.716613 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} err="failed to get container status \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": rpc error: code = NotFound desc = could not find container \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": container with ID starting with 32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.716632 4914 scope.go:117] "RemoveContainer" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.716959 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} err="failed to get container status \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": rpc error: code = NotFound desc = could not find container \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": container with ID starting with f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.716997 4914 scope.go:117] "RemoveContainer" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.717310 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} err="failed to get container status \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": rpc error: code = NotFound desc = could not find container \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": container with ID starting with e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.717329 4914 scope.go:117] "RemoveContainer" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.717674 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} err="failed to get container status \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": rpc error: code = NotFound desc = could not find container \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": container with ID starting with 15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.717722 4914 scope.go:117] "RemoveContainer" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.717982 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} err="failed to get container status \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": rpc error: code = NotFound desc = could not find container \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": container with ID starting with fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.718004 4914 scope.go:117] "RemoveContainer" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.718231 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} err="failed to get container status \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": rpc error: code = NotFound desc = could not find container \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": container with ID starting with 659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.718268 4914 scope.go:117] "RemoveContainer" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.718893 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} err="failed to get container status \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": rpc error: code = NotFound desc = could not find container \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": container with ID starting with 0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.718912 4914 scope.go:117] "RemoveContainer" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.719313 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} err="failed to get container status \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": rpc error: code = NotFound desc = could not find container \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": container with ID starting with a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.719351 4914 scope.go:117] "RemoveContainer" containerID="8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.719582 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9"} err="failed to get container status \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": rpc error: code = NotFound desc = could not find container \"8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9\": container with ID starting with 8dc598bfbfbd17e07520a6a62724f2b8ac284bab87334ceb2f857744758fe6a9 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.719600 4914 scope.go:117] "RemoveContainer" containerID="6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.719899 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09"} err="failed to get container status \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": rpc error: code = NotFound desc = could not find container \"6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09\": container with ID starting with 6683acf6bc44a1c2bb79f55885e443566fb5875eec2f412e0bccd5f1ec146c09 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.719943 4914 scope.go:117] "RemoveContainer" containerID="32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.722122 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f"} err="failed to get container status \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": rpc error: code = NotFound desc = could not find container \"32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f\": container with ID starting with 32a6917d8108138ec31d2627672bfd794b9670d20ce68c11741cf980e97ebd3f not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.722140 4914 scope.go:117] "RemoveContainer" containerID="f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.722656 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4"} err="failed to get container status \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": rpc error: code = NotFound desc = could not find container \"f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4\": container with ID starting with f1342050633c67fd25aacff0d4ff04290357899ff34c88db2dd9ecadd008ada4 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.722733 4914 scope.go:117] "RemoveContainer" containerID="e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.723238 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d"} err="failed to get container status \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": rpc error: code = NotFound desc = could not find container \"e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d\": container with ID starting with e1ddc6a9826c625ac15564d702d36bfe1a1dfd03aacb1db9a30d09c4f852c56d not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.723281 4914 scope.go:117] "RemoveContainer" containerID="15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.723614 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde"} err="failed to get container status \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": rpc error: code = NotFound desc = could not find container \"15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde\": container with ID starting with 15c1b9a9033cc9f8b4ced3eb8a9f43c50f9ec2e187fe3ddb25f3fd860e953cde not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.723655 4914 scope.go:117] "RemoveContainer" containerID="fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.724042 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024"} err="failed to get container status \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": rpc error: code = NotFound desc = could not find container \"fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024\": container with ID starting with fd887a728ad5fef21e8e0557c03aa576205c5eaba929db8f99a2ef1a2c779024 not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.724080 4914 scope.go:117] "RemoveContainer" containerID="659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.725441 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de"} err="failed to get container status \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": rpc error: code = NotFound desc = could not find container \"659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de\": container with ID starting with 659852d0ec106633b59b327a3d82c94e98788e7bf87a6fe8e79f98f4067c46de not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.725516 4914 scope.go:117] "RemoveContainer" containerID="0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.726230 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe"} err="failed to get container status \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": rpc error: code = NotFound desc = could not find container \"0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe\": container with ID starting with 0868a9aa07b0a800fa69e473dff344bae2db44c9254ea5abc9ca461cbbadb8fe not found: ID does not exist" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.726359 4914 scope.go:117] "RemoveContainer" containerID="a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5" Nov 22 14:36:45 crc kubenswrapper[4914]: I1122 14:36:45.726681 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5"} err="failed to get container status \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": rpc error: code = NotFound desc = could not find container \"a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5\": container with ID starting with a4a1d4144c60261bf13d48e16ceecd1acafd32bb2c4dc679862c63d0515c9fd5 not found: ID does not exist" Nov 22 14:36:46 crc kubenswrapper[4914]: I1122 14:36:46.486351 4914 generic.go:334] "Generic (PLEG): container finished" podID="c8e91974-4a2a-4b81-ae98-10c3ab849f57" containerID="a9cb6449c76cb632255b152a4531e7ecead1cd920190b3cddd173687c0f08005" exitCode=0 Nov 22 14:36:46 crc kubenswrapper[4914]: I1122 14:36:46.486394 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerDied","Data":"a9cb6449c76cb632255b152a4531e7ecead1cd920190b3cddd173687c0f08005"} Nov 22 14:36:46 crc kubenswrapper[4914]: I1122 14:36:46.486424 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"668c87e31cca07fbe153e80405eea04e48b46f7f2d2a22c2a311902d8690ad0e"} Nov 22 14:36:46 crc kubenswrapper[4914]: I1122 14:36:46.873896 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41ea1017-213b-4be4-87ae-d68655d58a12" path="/var/lib/kubelet/pods/41ea1017-213b-4be4-87ae-d68655d58a12/volumes" Nov 22 14:36:47 crc kubenswrapper[4914]: I1122 14:36:47.495453 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"43663bef3f926d63ed767ca37caf424ea08115e0ba4168d1d510a51edaedcde8"} Nov 22 14:36:47 crc kubenswrapper[4914]: I1122 14:36:47.495686 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"70306eca972cddccddb4b3a8cfb7b10c784913ef138683015ed6c63bdef410a8"} Nov 22 14:36:47 crc kubenswrapper[4914]: I1122 14:36:47.495699 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"e3a19fd41b29a30d83d1648dd1dd6348a9ff772be7c3babfb41142f781206476"} Nov 22 14:36:47 crc kubenswrapper[4914]: I1122 14:36:47.495710 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"d5ac5840372088a8c2d53feee73be08d32ebcfa03de777062b8f97df12175977"} Nov 22 14:36:47 crc kubenswrapper[4914]: I1122 14:36:47.495720 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"332728e92de249979fc15be67370fdf9909c4973bf81e38f4afb0eb8c36beb25"} Nov 22 14:36:47 crc kubenswrapper[4914]: I1122 14:36:47.495730 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"7046b7d1f91d244e58df8c2e34ec5539db38878f326eb994978388e7d169134b"} Nov 22 14:36:49 crc kubenswrapper[4914]: I1122 14:36:49.512537 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"63c3d5a22ae599331d2f1bb12fd8a8a431d88c7d61806853c13a8321b91a53fa"} Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.042566 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.042997 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.043064 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.043926 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8bb85fe0e7b9128d7e9a8dfe78ff109dee5730d63a76ca28de3d0e288cea71db"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.044024 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://8bb85fe0e7b9128d7e9a8dfe78ff109dee5730d63a76ca28de3d0e288cea71db" gracePeriod=600 Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.529571 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="8bb85fe0e7b9128d7e9a8dfe78ff109dee5730d63a76ca28de3d0e288cea71db" exitCode=0 Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.529669 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"8bb85fe0e7b9128d7e9a8dfe78ff109dee5730d63a76ca28de3d0e288cea71db"} Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.529971 4914 scope.go:117] "RemoveContainer" containerID="295e4be82247193043ad744eafabe554329435763a26422ff3518c50082e23e2" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.535187 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" event={"ID":"c8e91974-4a2a-4b81-ae98-10c3ab849f57","Type":"ContainerStarted","Data":"df1b97c3b2d1dcb87bfa830d20dc026bfe63fcb59ad18932a542e7733d399edf"} Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.535559 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.535595 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.565370 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" podStartSLOduration=7.565352906 podStartE2EDuration="7.565352906s" podCreationTimestamp="2025-11-22 14:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:36:52.562663531 +0000 UTC m=+578.371233212" watchObservedRunningTime="2025-11-22 14:36:52.565352906 +0000 UTC m=+578.373922577" Nov 22 14:36:52 crc kubenswrapper[4914]: I1122 14:36:52.566302 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:53 crc kubenswrapper[4914]: I1122 14:36:53.549081 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"448c460df7aa9c496e7a50b996cc4b57b27198da3e1c4c2cbdb4f3896aa006fa"} Nov 22 14:36:53 crc kubenswrapper[4914]: I1122 14:36:53.549455 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:53 crc kubenswrapper[4914]: I1122 14:36:53.595613 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:36:56 crc kubenswrapper[4914]: I1122 14:36:56.910945 4914 scope.go:117] "RemoveContainer" containerID="bdb8ada03e70d054a4dfb6fa8af17dd17837b11a178753fb893b8fa6403e4862" Nov 22 14:36:56 crc kubenswrapper[4914]: E1122 14:36:56.911575 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8hpjt_openshift-multus(b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed)\"" pod="openshift-multus/multus-8hpjt" podUID="b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed" Nov 22 14:37:09 crc kubenswrapper[4914]: I1122 14:37:09.868485 4914 scope.go:117] "RemoveContainer" containerID="bdb8ada03e70d054a4dfb6fa8af17dd17837b11a178753fb893b8fa6403e4862" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.508984 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr"] Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.510403 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.513212 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.525075 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr"] Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.635894 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.635949 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.636037 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72gp6\" (UniqueName: \"kubernetes.io/projected/a165a22d-31b5-46dc-835a-7e735a35da62-kube-api-access-72gp6\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.650462 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/2.log" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.651236 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/1.log" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.651280 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8hpjt" event={"ID":"b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed","Type":"ContainerStarted","Data":"5f6d5fb8e3754e48ff4918226d4a80647101f15afe77aad578be1fa0ff1ce4be"} Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.737047 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.737217 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72gp6\" (UniqueName: \"kubernetes.io/projected/a165a22d-31b5-46dc-835a-7e735a35da62-kube-api-access-72gp6\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.737308 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.738043 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.738403 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.763137 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72gp6\" (UniqueName: \"kubernetes.io/projected/a165a22d-31b5-46dc-835a-7e735a35da62-kube-api-access-72gp6\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: I1122 14:37:10.824370 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: E1122 14:37:10.854125 4914 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_openshift-marketplace_a165a22d-31b5-46dc-835a-7e735a35da62_0(eb10ff4545eada3723ffa7e7daf5896034738bccb8e6f9b9bbd2ea861c5851b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 22 14:37:10 crc kubenswrapper[4914]: E1122 14:37:10.854609 4914 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_openshift-marketplace_a165a22d-31b5-46dc-835a-7e735a35da62_0(eb10ff4545eada3723ffa7e7daf5896034738bccb8e6f9b9bbd2ea861c5851b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: E1122 14:37:10.854659 4914 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_openshift-marketplace_a165a22d-31b5-46dc-835a-7e735a35da62_0(eb10ff4545eada3723ffa7e7daf5896034738bccb8e6f9b9bbd2ea861c5851b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:10 crc kubenswrapper[4914]: E1122 14:37:10.854784 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_openshift-marketplace(a165a22d-31b5-46dc-835a-7e735a35da62)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_openshift-marketplace(a165a22d-31b5-46dc-835a-7e735a35da62)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_openshift-marketplace_a165a22d-31b5-46dc-835a-7e735a35da62_0(eb10ff4545eada3723ffa7e7daf5896034738bccb8e6f9b9bbd2ea861c5851b2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" Nov 22 14:37:11 crc kubenswrapper[4914]: I1122 14:37:11.660642 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:11 crc kubenswrapper[4914]: I1122 14:37:11.661861 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:12 crc kubenswrapper[4914]: I1122 14:37:12.055596 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr"] Nov 22 14:37:12 crc kubenswrapper[4914]: I1122 14:37:12.667475 4914 generic.go:334] "Generic (PLEG): container finished" podID="a165a22d-31b5-46dc-835a-7e735a35da62" containerID="8d01c0c835ce1d9835007a4b31d0b72d5c98a2988162f6780cae56f056758b26" exitCode=0 Nov 22 14:37:12 crc kubenswrapper[4914]: I1122 14:37:12.667530 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" event={"ID":"a165a22d-31b5-46dc-835a-7e735a35da62","Type":"ContainerDied","Data":"8d01c0c835ce1d9835007a4b31d0b72d5c98a2988162f6780cae56f056758b26"} Nov 22 14:37:12 crc kubenswrapper[4914]: I1122 14:37:12.667702 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" event={"ID":"a165a22d-31b5-46dc-835a-7e735a35da62","Type":"ContainerStarted","Data":"95bd6dcb0303f13bed1c73ecb3403b8f951f1dfdd3681817f58da8ab90bbd9e2"} Nov 22 14:37:12 crc kubenswrapper[4914]: I1122 14:37:12.670018 4914 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 14:37:14 crc kubenswrapper[4914]: I1122 14:37:14.682308 4914 generic.go:334] "Generic (PLEG): container finished" podID="a165a22d-31b5-46dc-835a-7e735a35da62" containerID="60e168cdc4ac7c4c1bf0ff51c8c3b99c6779a44392340d2e0597ea1db2d1c7a1" exitCode=0 Nov 22 14:37:14 crc kubenswrapper[4914]: I1122 14:37:14.682385 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" event={"ID":"a165a22d-31b5-46dc-835a-7e735a35da62","Type":"ContainerDied","Data":"60e168cdc4ac7c4c1bf0ff51c8c3b99c6779a44392340d2e0597ea1db2d1c7a1"} Nov 22 14:37:15 crc kubenswrapper[4914]: I1122 14:37:15.078560 4914 scope.go:117] "RemoveContainer" containerID="dd1e996067f18aebe257aec1d0ddd06a4444d46d264fd1898740e9dfbc16e879" Nov 22 14:37:15 crc kubenswrapper[4914]: I1122 14:37:15.648955 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9v8j5" Nov 22 14:37:15 crc kubenswrapper[4914]: I1122 14:37:15.698310 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8hpjt_b1f94050-7d2c-4a11-9bb6-ce9c9cdaabed/kube-multus/2.log" Nov 22 14:37:15 crc kubenswrapper[4914]: I1122 14:37:15.700671 4914 generic.go:334] "Generic (PLEG): container finished" podID="a165a22d-31b5-46dc-835a-7e735a35da62" containerID="0cd24ecbf33f6f350a9e8acea83c844a44247eede5eb5f99ecdb9e04b162a6d7" exitCode=0 Nov 22 14:37:15 crc kubenswrapper[4914]: I1122 14:37:15.700712 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" event={"ID":"a165a22d-31b5-46dc-835a-7e735a35da62","Type":"ContainerDied","Data":"0cd24ecbf33f6f350a9e8acea83c844a44247eede5eb5f99ecdb9e04b162a6d7"} Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.017164 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.120924 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-util\") pod \"a165a22d-31b5-46dc-835a-7e735a35da62\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.121013 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-bundle\") pod \"a165a22d-31b5-46dc-835a-7e735a35da62\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.121218 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72gp6\" (UniqueName: \"kubernetes.io/projected/a165a22d-31b5-46dc-835a-7e735a35da62-kube-api-access-72gp6\") pod \"a165a22d-31b5-46dc-835a-7e735a35da62\" (UID: \"a165a22d-31b5-46dc-835a-7e735a35da62\") " Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.122550 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-bundle" (OuterVolumeSpecName: "bundle") pod "a165a22d-31b5-46dc-835a-7e735a35da62" (UID: "a165a22d-31b5-46dc-835a-7e735a35da62"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.128221 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a165a22d-31b5-46dc-835a-7e735a35da62-kube-api-access-72gp6" (OuterVolumeSpecName: "kube-api-access-72gp6") pod "a165a22d-31b5-46dc-835a-7e735a35da62" (UID: "a165a22d-31b5-46dc-835a-7e735a35da62"). InnerVolumeSpecName "kube-api-access-72gp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.152715 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-util" (OuterVolumeSpecName: "util") pod "a165a22d-31b5-46dc-835a-7e735a35da62" (UID: "a165a22d-31b5-46dc-835a-7e735a35da62"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.223240 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72gp6\" (UniqueName: \"kubernetes.io/projected/a165a22d-31b5-46dc-835a-7e735a35da62-kube-api-access-72gp6\") on node \"crc\" DevicePath \"\"" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.223294 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.223318 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a165a22d-31b5-46dc-835a-7e735a35da62-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.716864 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" event={"ID":"a165a22d-31b5-46dc-835a-7e735a35da62","Type":"ContainerDied","Data":"95bd6dcb0303f13bed1c73ecb3403b8f951f1dfdd3681817f58da8ab90bbd9e2"} Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.716949 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95bd6dcb0303f13bed1c73ecb3403b8f951f1dfdd3681817f58da8ab90bbd9e2" Nov 22 14:37:17 crc kubenswrapper[4914]: I1122 14:37:17.717011 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.700380 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm"] Nov 22 14:37:28 crc kubenswrapper[4914]: E1122 14:37:28.700959 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="pull" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.700969 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="pull" Nov 22 14:37:28 crc kubenswrapper[4914]: E1122 14:37:28.700984 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="util" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.700991 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="util" Nov 22 14:37:28 crc kubenswrapper[4914]: E1122 14:37:28.701001 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="extract" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.701007 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="extract" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.701101 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="a165a22d-31b5-46dc-835a-7e735a35da62" containerName="extract" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.701432 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.703273 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.703315 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.703338 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.703419 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.703689 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-mvljk" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.717360 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm"] Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.775393 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7addd0c3-a053-491c-b6a4-3a85cc0263ed-webhook-cert\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.775438 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7addd0c3-a053-491c-b6a4-3a85cc0263ed-apiservice-cert\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.775488 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llvtl\" (UniqueName: \"kubernetes.io/projected/7addd0c3-a053-491c-b6a4-3a85cc0263ed-kube-api-access-llvtl\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.876765 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llvtl\" (UniqueName: \"kubernetes.io/projected/7addd0c3-a053-491c-b6a4-3a85cc0263ed-kube-api-access-llvtl\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.877012 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7addd0c3-a053-491c-b6a4-3a85cc0263ed-webhook-cert\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.877096 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7addd0c3-a053-491c-b6a4-3a85cc0263ed-apiservice-cert\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.882143 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7addd0c3-a053-491c-b6a4-3a85cc0263ed-webhook-cert\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.887798 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7addd0c3-a053-491c-b6a4-3a85cc0263ed-apiservice-cert\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:28 crc kubenswrapper[4914]: I1122 14:37:28.898389 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llvtl\" (UniqueName: \"kubernetes.io/projected/7addd0c3-a053-491c-b6a4-3a85cc0263ed-kube-api-access-llvtl\") pod \"metallb-operator-controller-manager-5cb954cd64-99pbm\" (UID: \"7addd0c3-a053-491c-b6a4-3a85cc0263ed\") " pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.015548 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.036008 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5496757488-29bhv"] Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.036997 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.038775 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.039110 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.039121 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-257td" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.048941 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5496757488-29bhv"] Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.081748 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f9c1310-eb65-4286-8466-e671dbac0ec1-webhook-cert\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.081826 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f9c1310-eb65-4286-8466-e671dbac0ec1-apiservice-cert\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.081904 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jwsk\" (UniqueName: \"kubernetes.io/projected/7f9c1310-eb65-4286-8466-e671dbac0ec1-kube-api-access-4jwsk\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.196313 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f9c1310-eb65-4286-8466-e671dbac0ec1-webhook-cert\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.196591 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f9c1310-eb65-4286-8466-e671dbac0ec1-apiservice-cert\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.196616 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jwsk\" (UniqueName: \"kubernetes.io/projected/7f9c1310-eb65-4286-8466-e671dbac0ec1-kube-api-access-4jwsk\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.202582 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f9c1310-eb65-4286-8466-e671dbac0ec1-apiservice-cert\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.215674 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f9c1310-eb65-4286-8466-e671dbac0ec1-webhook-cert\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.222517 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jwsk\" (UniqueName: \"kubernetes.io/projected/7f9c1310-eb65-4286-8466-e671dbac0ec1-kube-api-access-4jwsk\") pod \"metallb-operator-webhook-server-5496757488-29bhv\" (UID: \"7f9c1310-eb65-4286-8466-e671dbac0ec1\") " pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.291649 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm"] Nov 22 14:37:29 crc kubenswrapper[4914]: W1122 14:37:29.303029 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7addd0c3_a053_491c_b6a4_3a85cc0263ed.slice/crio-44a5b90a112deb943b19054236b57ba9562882853771e0248e1e48cb6bbb681d WatchSource:0}: Error finding container 44a5b90a112deb943b19054236b57ba9562882853771e0248e1e48cb6bbb681d: Status 404 returned error can't find the container with id 44a5b90a112deb943b19054236b57ba9562882853771e0248e1e48cb6bbb681d Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.379480 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.549172 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5496757488-29bhv"] Nov 22 14:37:29 crc kubenswrapper[4914]: W1122 14:37:29.557302 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f9c1310_eb65_4286_8466_e671dbac0ec1.slice/crio-51585c28bcf2d9b8b8836eb38da048c78ae0d39f83dee3e20d7b3c8cc693db8b WatchSource:0}: Error finding container 51585c28bcf2d9b8b8836eb38da048c78ae0d39f83dee3e20d7b3c8cc693db8b: Status 404 returned error can't find the container with id 51585c28bcf2d9b8b8836eb38da048c78ae0d39f83dee3e20d7b3c8cc693db8b Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.790341 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" event={"ID":"7addd0c3-a053-491c-b6a4-3a85cc0263ed","Type":"ContainerStarted","Data":"44a5b90a112deb943b19054236b57ba9562882853771e0248e1e48cb6bbb681d"} Nov 22 14:37:29 crc kubenswrapper[4914]: I1122 14:37:29.792518 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" event={"ID":"7f9c1310-eb65-4286-8466-e671dbac0ec1","Type":"ContainerStarted","Data":"51585c28bcf2d9b8b8836eb38da048c78ae0d39f83dee3e20d7b3c8cc693db8b"} Nov 22 14:37:32 crc kubenswrapper[4914]: I1122 14:37:32.898947 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" event={"ID":"7addd0c3-a053-491c-b6a4-3a85cc0263ed","Type":"ContainerStarted","Data":"b8e8ebdd8b1639b2f5739e04c0b01620a08c3a550061ad218afdb3a49735908c"} Nov 22 14:37:32 crc kubenswrapper[4914]: I1122 14:37:32.899410 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:37:33 crc kubenswrapper[4914]: I1122 14:37:33.890560 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" event={"ID":"7f9c1310-eb65-4286-8466-e671dbac0ec1","Type":"ContainerStarted","Data":"de6cef406b1671b3d71b3e0bc38e1eec7c8dd4df51814b2c69758381fbaf8b0f"} Nov 22 14:37:33 crc kubenswrapper[4914]: I1122 14:37:33.911817 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" podStartSLOduration=3.382475972 podStartE2EDuration="5.911795638s" podCreationTimestamp="2025-11-22 14:37:28 +0000 UTC" firstStartedPulling="2025-11-22 14:37:29.306401 +0000 UTC m=+615.114970661" lastFinishedPulling="2025-11-22 14:37:31.835720666 +0000 UTC m=+617.644290327" observedRunningTime="2025-11-22 14:37:32.927095359 +0000 UTC m=+618.735665020" watchObservedRunningTime="2025-11-22 14:37:33.911795638 +0000 UTC m=+619.720365309" Nov 22 14:37:33 crc kubenswrapper[4914]: I1122 14:37:33.912114 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" podStartSLOduration=0.787165753 podStartE2EDuration="4.912109676s" podCreationTimestamp="2025-11-22 14:37:29 +0000 UTC" firstStartedPulling="2025-11-22 14:37:29.559735836 +0000 UTC m=+615.368305487" lastFinishedPulling="2025-11-22 14:37:33.684679749 +0000 UTC m=+619.493249410" observedRunningTime="2025-11-22 14:37:33.906454828 +0000 UTC m=+619.715024509" watchObservedRunningTime="2025-11-22 14:37:33.912109676 +0000 UTC m=+619.720679347" Nov 22 14:37:34 crc kubenswrapper[4914]: I1122 14:37:34.896661 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:37:49 crc kubenswrapper[4914]: I1122 14:37:49.383024 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5496757488-29bhv" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.018140 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5cb954cd64-99pbm" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.939972 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv"] Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.941154 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.945496 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-vkmkt"] Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.946091 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.946595 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4kg9s" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.949011 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.951188 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.951477 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 22 14:38:09 crc kubenswrapper[4914]: I1122 14:38:09.988372 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv"] Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027610 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-startup\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027669 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-conf\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027701 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7-cert\") pod \"frr-k8s-webhook-server-6998585d5-vg5hv\" (UID: \"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027727 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-sockets\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027761 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027799 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czcl8\" (UniqueName: \"kubernetes.io/projected/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-kube-api-access-czcl8\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027842 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fhq9\" (UniqueName: \"kubernetes.io/projected/e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7-kube-api-access-4fhq9\") pod \"frr-k8s-webhook-server-6998585d5-vg5hv\" (UID: \"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027931 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics-certs\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.027964 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-reloader\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.048277 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ppqw8"] Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.049291 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.052287 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.052746 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.053108 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.053133 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-zhm6r" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.069119 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-58dgb"] Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.069905 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: W1122 14:38:10.072464 4914 reflector.go:561] object-"metallb-system"/"controller-certs-secret": failed to list *v1.Secret: secrets "controller-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.072495 4914 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"controller-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"controller-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.092947 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-58dgb"] Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129330 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129385 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czcl8\" (UniqueName: \"kubernetes.io/projected/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-kube-api-access-czcl8\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129415 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-cert\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129437 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-metrics-certs\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129468 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fhq9\" (UniqueName: \"kubernetes.io/projected/e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7-kube-api-access-4fhq9\") pod \"frr-k8s-webhook-server-6998585d5-vg5hv\" (UID: \"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129491 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics-certs\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129508 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129533 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-reloader\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129549 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-metallb-excludel2\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129566 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhhpn\" (UniqueName: \"kubernetes.io/projected/5a4d4454-920a-4ec8-a619-13f217a6d027-kube-api-access-mhhpn\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129590 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbsv7\" (UniqueName: \"kubernetes.io/projected/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-kube-api-access-zbsv7\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129607 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-startup\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129626 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-conf\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129647 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7-cert\") pod \"frr-k8s-webhook-server-6998585d5-vg5hv\" (UID: \"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129663 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-sockets\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129703 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-metrics-certs\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.129736 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.130293 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-reloader\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.130352 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-conf\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.130545 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-sockets\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.130665 4914 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.130717 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics-certs podName:6b1e6db9-848b-48a6-bfc4-053961e4ecb3 nodeName:}" failed. No retries permitted until 2025-11-22 14:38:10.63070394 +0000 UTC m=+656.439273601 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics-certs") pod "frr-k8s-vkmkt" (UID: "6b1e6db9-848b-48a6-bfc4-053961e4ecb3") : secret "frr-k8s-certs-secret" not found Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.130940 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-frr-startup\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.144406 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fhq9\" (UniqueName: \"kubernetes.io/projected/e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7-kube-api-access-4fhq9\") pod \"frr-k8s-webhook-server-6998585d5-vg5hv\" (UID: \"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.147517 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czcl8\" (UniqueName: \"kubernetes.io/projected/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-kube-api-access-czcl8\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.149531 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7-cert\") pod \"frr-k8s-webhook-server-6998585d5-vg5hv\" (UID: \"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230489 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-metrics-certs\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230539 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-cert\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230559 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-metrics-certs\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230596 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230621 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-metallb-excludel2\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230637 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhhpn\" (UniqueName: \"kubernetes.io/projected/5a4d4454-920a-4ec8-a619-13f217a6d027-kube-api-access-mhhpn\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.230663 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbsv7\" (UniqueName: \"kubernetes.io/projected/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-kube-api-access-zbsv7\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.230774 4914 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.230853 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist podName:499810cd-69d9-49de-aa7f-a7b20ba8f7b5 nodeName:}" failed. No retries permitted until 2025-11-22 14:38:10.730835999 +0000 UTC m=+656.539405660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist") pod "speaker-ppqw8" (UID: "499810cd-69d9-49de-aa7f-a7b20ba8f7b5") : secret "metallb-memberlist" not found Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.231547 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-metallb-excludel2\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.236156 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-metrics-certs\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.236389 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-cert\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.252510 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhhpn\" (UniqueName: \"kubernetes.io/projected/5a4d4454-920a-4ec8-a619-13f217a6d027-kube-api-access-mhhpn\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.256456 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbsv7\" (UniqueName: \"kubernetes.io/projected/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-kube-api-access-zbsv7\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.288848 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.636427 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics-certs\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.643310 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b1e6db9-848b-48a6-bfc4-053961e4ecb3-metrics-certs\") pod \"frr-k8s-vkmkt\" (UID: \"6b1e6db9-848b-48a6-bfc4-053961e4ecb3\") " pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.715357 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv"] Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.738202 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.738345 4914 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 14:38:10 crc kubenswrapper[4914]: E1122 14:38:10.738442 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist podName:499810cd-69d9-49de-aa7f-a7b20ba8f7b5 nodeName:}" failed. No retries permitted until 2025-11-22 14:38:11.738418027 +0000 UTC m=+657.546987698 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist") pod "speaker-ppqw8" (UID: "499810cd-69d9-49de-aa7f-a7b20ba8f7b5") : secret "metallb-memberlist" not found Nov 22 14:38:10 crc kubenswrapper[4914]: I1122 14:38:10.901339 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.106439 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"0efce8f031c007cb042fb5b94066d05e82738936696d91c58990332cbf4020d0"} Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.110616 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" event={"ID":"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7","Type":"ContainerStarted","Data":"81573cf82a10f9019704300381df44ff57cdfca3188e5a0216082c4706fec8c5"} Nov 22 14:38:11 crc kubenswrapper[4914]: E1122 14:38:11.231392 4914 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: failed to sync secret cache: timed out waiting for the condition Nov 22 14:38:11 crc kubenswrapper[4914]: E1122 14:38:11.231527 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-metrics-certs podName:5a4d4454-920a-4ec8-a619-13f217a6d027 nodeName:}" failed. No retries permitted until 2025-11-22 14:38:11.731482034 +0000 UTC m=+657.540051735 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-metrics-certs") pod "controller-6c7b4b5f48-58dgb" (UID: "5a4d4454-920a-4ec8-a619-13f217a6d027") : failed to sync secret cache: timed out waiting for the condition Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.673993 4914 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.752622 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-metrics-certs\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.752765 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:11 crc kubenswrapper[4914]: E1122 14:38:11.753029 4914 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 22 14:38:11 crc kubenswrapper[4914]: E1122 14:38:11.753128 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist podName:499810cd-69d9-49de-aa7f-a7b20ba8f7b5 nodeName:}" failed. No retries permitted until 2025-11-22 14:38:13.753105494 +0000 UTC m=+659.561675165 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist") pod "speaker-ppqw8" (UID: "499810cd-69d9-49de-aa7f-a7b20ba8f7b5") : secret "metallb-memberlist" not found Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.761694 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5a4d4454-920a-4ec8-a619-13f217a6d027-metrics-certs\") pod \"controller-6c7b4b5f48-58dgb\" (UID: \"5a4d4454-920a-4ec8-a619-13f217a6d027\") " pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:11 crc kubenswrapper[4914]: I1122 14:38:11.882135 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:12 crc kubenswrapper[4914]: I1122 14:38:12.081161 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-58dgb"] Nov 22 14:38:13 crc kubenswrapper[4914]: I1122 14:38:13.123278 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-58dgb" event={"ID":"5a4d4454-920a-4ec8-a619-13f217a6d027","Type":"ContainerStarted","Data":"6a88b5af273ba2776f47eef6f4505815d51b095eab10ef1511b4b5e9f3dcb1de"} Nov 22 14:38:13 crc kubenswrapper[4914]: I1122 14:38:13.123542 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-58dgb" event={"ID":"5a4d4454-920a-4ec8-a619-13f217a6d027","Type":"ContainerStarted","Data":"c1dc14e470ab36383fb03e48ae2427e53150fed049eba049b3a74907cb6066ba"} Nov 22 14:38:13 crc kubenswrapper[4914]: I1122 14:38:13.811084 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:13 crc kubenswrapper[4914]: I1122 14:38:13.820629 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/499810cd-69d9-49de-aa7f-a7b20ba8f7b5-memberlist\") pod \"speaker-ppqw8\" (UID: \"499810cd-69d9-49de-aa7f-a7b20ba8f7b5\") " pod="metallb-system/speaker-ppqw8" Nov 22 14:38:13 crc kubenswrapper[4914]: I1122 14:38:13.961625 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ppqw8" Nov 22 14:38:14 crc kubenswrapper[4914]: W1122 14:38:14.026623 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod499810cd_69d9_49de_aa7f_a7b20ba8f7b5.slice/crio-059d253255bc201510c2d415e67bcc606545eba4306e443b2fa1fe7c0f0e1de0 WatchSource:0}: Error finding container 059d253255bc201510c2d415e67bcc606545eba4306e443b2fa1fe7c0f0e1de0: Status 404 returned error can't find the container with id 059d253255bc201510c2d415e67bcc606545eba4306e443b2fa1fe7c0f0e1de0 Nov 22 14:38:14 crc kubenswrapper[4914]: I1122 14:38:14.145560 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ppqw8" event={"ID":"499810cd-69d9-49de-aa7f-a7b20ba8f7b5","Type":"ContainerStarted","Data":"059d253255bc201510c2d415e67bcc606545eba4306e443b2fa1fe7c0f0e1de0"} Nov 22 14:38:15 crc kubenswrapper[4914]: I1122 14:38:15.158475 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ppqw8" event={"ID":"499810cd-69d9-49de-aa7f-a7b20ba8f7b5","Type":"ContainerStarted","Data":"fc52585ceab1f76c78f26977f4135553e207659934f983eac3461f1bc9eb9443"} Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.174923 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ppqw8" event={"ID":"499810cd-69d9-49de-aa7f-a7b20ba8f7b5","Type":"ContainerStarted","Data":"26d6b2ed6cb5027423be42d164dbf01d36a62b53e4a69a471c3090d65636be64"} Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.175307 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ppqw8" Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.176656 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-58dgb" event={"ID":"5a4d4454-920a-4ec8-a619-13f217a6d027","Type":"ContainerStarted","Data":"b584cf4342d4153bce3130a51c2a08b2212f33f56268483d3ef4a3cfd26c472a"} Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.176734 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.178505 4914 generic.go:334] "Generic (PLEG): container finished" podID="6b1e6db9-848b-48a6-bfc4-053961e4ecb3" containerID="7d2f6d277463a054b8c7933d7fc1851f3cf13e92d7756a9cc7615604add886f8" exitCode=0 Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.178559 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerDied","Data":"7d2f6d277463a054b8c7933d7fc1851f3cf13e92d7756a9cc7615604add886f8"} Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.180074 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" event={"ID":"e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7","Type":"ContainerStarted","Data":"0accec703c321ea2219902adf69d10f8da24c7e8242e81ae963532270f29a1d9"} Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.180242 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.203446 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ppqw8" podStartSLOduration=4.8846958879999995 podStartE2EDuration="8.203428139s" podCreationTimestamp="2025-11-22 14:38:10 +0000 UTC" firstStartedPulling="2025-11-22 14:38:14.367369512 +0000 UTC m=+660.175939173" lastFinishedPulling="2025-11-22 14:38:17.686101753 +0000 UTC m=+663.494671424" observedRunningTime="2025-11-22 14:38:18.201496623 +0000 UTC m=+664.010066344" watchObservedRunningTime="2025-11-22 14:38:18.203428139 +0000 UTC m=+664.011997800" Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.222964 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-58dgb" podStartSLOduration=2.943204126 podStartE2EDuration="8.222948003s" podCreationTimestamp="2025-11-22 14:38:10 +0000 UTC" firstStartedPulling="2025-11-22 14:38:12.397101252 +0000 UTC m=+658.205670913" lastFinishedPulling="2025-11-22 14:38:17.676845119 +0000 UTC m=+663.485414790" observedRunningTime="2025-11-22 14:38:18.21704814 +0000 UTC m=+664.025617801" watchObservedRunningTime="2025-11-22 14:38:18.222948003 +0000 UTC m=+664.031517664" Nov 22 14:38:18 crc kubenswrapper[4914]: I1122 14:38:18.242454 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" podStartSLOduration=2.279992101 podStartE2EDuration="9.242429225s" podCreationTimestamp="2025-11-22 14:38:09 +0000 UTC" firstStartedPulling="2025-11-22 14:38:10.721332143 +0000 UTC m=+656.529901844" lastFinishedPulling="2025-11-22 14:38:17.683769297 +0000 UTC m=+663.492338968" observedRunningTime="2025-11-22 14:38:18.239220377 +0000 UTC m=+664.047790098" watchObservedRunningTime="2025-11-22 14:38:18.242429225 +0000 UTC m=+664.050998916" Nov 22 14:38:19 crc kubenswrapper[4914]: I1122 14:38:19.191580 4914 generic.go:334] "Generic (PLEG): container finished" podID="6b1e6db9-848b-48a6-bfc4-053961e4ecb3" containerID="5b55922da7b59e371d7ee17601a99f710da504898862c00b3499b74cf41d783c" exitCode=0 Nov 22 14:38:19 crc kubenswrapper[4914]: I1122 14:38:19.191668 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerDied","Data":"5b55922da7b59e371d7ee17601a99f710da504898862c00b3499b74cf41d783c"} Nov 22 14:38:20 crc kubenswrapper[4914]: I1122 14:38:20.201689 4914 generic.go:334] "Generic (PLEG): container finished" podID="6b1e6db9-848b-48a6-bfc4-053961e4ecb3" containerID="ec2617b17be0887aa38e9f7822de01d91c43f4e5c0877741a968b7440ccf67e6" exitCode=0 Nov 22 14:38:20 crc kubenswrapper[4914]: I1122 14:38:20.201747 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerDied","Data":"ec2617b17be0887aa38e9f7822de01d91c43f4e5c0877741a968b7440ccf67e6"} Nov 22 14:38:21 crc kubenswrapper[4914]: I1122 14:38:21.213060 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"3388b2f851112dd71c557184b226a005b47ff3076ad9268674ada0ac8234ac12"} Nov 22 14:38:21 crc kubenswrapper[4914]: I1122 14:38:21.213442 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"d00d2ce57bf6ca56338dedd9d821d5825f6821dfbdec8cea0e2d364f6321e195"} Nov 22 14:38:21 crc kubenswrapper[4914]: I1122 14:38:21.213454 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"7db4e8a4588ae655905b08f0639f6b6b869d325877548a140571103bc3fa8062"} Nov 22 14:38:22 crc kubenswrapper[4914]: I1122 14:38:22.227650 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"cf15a67d11d345d28e7be0498d69e94d50aee4e1b320d5974f0de6167b12d690"} Nov 22 14:38:22 crc kubenswrapper[4914]: I1122 14:38:22.228320 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:22 crc kubenswrapper[4914]: I1122 14:38:22.228349 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"929482493dfad1fb5d7d025800c01fee8590597968f36c78c9a9b35f1cfb43c7"} Nov 22 14:38:22 crc kubenswrapper[4914]: I1122 14:38:22.228376 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vkmkt" event={"ID":"6b1e6db9-848b-48a6-bfc4-053961e4ecb3","Type":"ContainerStarted","Data":"f1f8d687f68125d6f49e15f13d73e960558c2eff0dcafc4fa9f78fe9e45b8885"} Nov 22 14:38:22 crc kubenswrapper[4914]: I1122 14:38:22.263866 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-vkmkt" podStartSLOduration=6.532288923 podStartE2EDuration="13.263841327s" podCreationTimestamp="2025-11-22 14:38:09 +0000 UTC" firstStartedPulling="2025-11-22 14:38:11.031589957 +0000 UTC m=+656.840159628" lastFinishedPulling="2025-11-22 14:38:17.763142331 +0000 UTC m=+663.571712032" observedRunningTime="2025-11-22 14:38:22.254832728 +0000 UTC m=+668.063402399" watchObservedRunningTime="2025-11-22 14:38:22.263841327 +0000 UTC m=+668.072411018" Nov 22 14:38:25 crc kubenswrapper[4914]: I1122 14:38:25.901697 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:25 crc kubenswrapper[4914]: I1122 14:38:25.950534 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:30 crc kubenswrapper[4914]: I1122 14:38:30.297250 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-vg5hv" Nov 22 14:38:30 crc kubenswrapper[4914]: I1122 14:38:30.906677 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-vkmkt" Nov 22 14:38:31 crc kubenswrapper[4914]: I1122 14:38:31.887654 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-58dgb" Nov 22 14:38:33 crc kubenswrapper[4914]: I1122 14:38:33.965902 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ppqw8" Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.871110 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-4rkdv"] Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.873524 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.885016 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.885283 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-4rkdv"] Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.885547 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.886161 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-zdtmz" Nov 22 14:38:39 crc kubenswrapper[4914]: I1122 14:38:39.964371 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bwwl\" (UniqueName: \"kubernetes.io/projected/b1d67ef8-0905-4ec7-8b01-75d765f1f9f3-kube-api-access-4bwwl\") pod \"mariadb-operator-index-4rkdv\" (UID: \"b1d67ef8-0905-4ec7-8b01-75d765f1f9f3\") " pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:40 crc kubenswrapper[4914]: I1122 14:38:40.065767 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bwwl\" (UniqueName: \"kubernetes.io/projected/b1d67ef8-0905-4ec7-8b01-75d765f1f9f3-kube-api-access-4bwwl\") pod \"mariadb-operator-index-4rkdv\" (UID: \"b1d67ef8-0905-4ec7-8b01-75d765f1f9f3\") " pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:40 crc kubenswrapper[4914]: I1122 14:38:40.083169 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bwwl\" (UniqueName: \"kubernetes.io/projected/b1d67ef8-0905-4ec7-8b01-75d765f1f9f3-kube-api-access-4bwwl\") pod \"mariadb-operator-index-4rkdv\" (UID: \"b1d67ef8-0905-4ec7-8b01-75d765f1f9f3\") " pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:40 crc kubenswrapper[4914]: I1122 14:38:40.209750 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:40 crc kubenswrapper[4914]: I1122 14:38:40.440611 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-4rkdv"] Nov 22 14:38:41 crc kubenswrapper[4914]: I1122 14:38:41.344941 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-4rkdv" event={"ID":"b1d67ef8-0905-4ec7-8b01-75d765f1f9f3","Type":"ContainerStarted","Data":"e4ccfa60ff73aa612223ad4d057423564100273ce800c0412318b9ac21a10cd1"} Nov 22 14:38:42 crc kubenswrapper[4914]: I1122 14:38:42.352174 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-4rkdv" event={"ID":"b1d67ef8-0905-4ec7-8b01-75d765f1f9f3","Type":"ContainerStarted","Data":"06005bae622ea4622189b9cf855f875724da48ff0b23896cec402dbed13eb94a"} Nov 22 14:38:42 crc kubenswrapper[4914]: I1122 14:38:42.379296 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-4rkdv" podStartSLOduration=2.403421984 podStartE2EDuration="3.379272389s" podCreationTimestamp="2025-11-22 14:38:39 +0000 UTC" firstStartedPulling="2025-11-22 14:38:40.462601608 +0000 UTC m=+686.271171269" lastFinishedPulling="2025-11-22 14:38:41.438452013 +0000 UTC m=+687.247021674" observedRunningTime="2025-11-22 14:38:42.374687847 +0000 UTC m=+688.183257578" watchObservedRunningTime="2025-11-22 14:38:42.379272389 +0000 UTC m=+688.187842080" Nov 22 14:38:50 crc kubenswrapper[4914]: I1122 14:38:50.210661 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:50 crc kubenswrapper[4914]: I1122 14:38:50.211356 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:50 crc kubenswrapper[4914]: I1122 14:38:50.247571 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:50 crc kubenswrapper[4914]: I1122 14:38:50.441072 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-4rkdv" Nov 22 14:38:52 crc kubenswrapper[4914]: I1122 14:38:52.042290 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:38:52 crc kubenswrapper[4914]: I1122 14:38:52.042616 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.856496 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl"] Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.858083 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.860955 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l5bmb" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.868444 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl"] Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.890978 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cb9t\" (UniqueName: \"kubernetes.io/projected/d0a62272-b520-435c-a59e-783429f5638b-kube-api-access-2cb9t\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.891042 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-bundle\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.891103 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-util\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.991710 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cb9t\" (UniqueName: \"kubernetes.io/projected/d0a62272-b520-435c-a59e-783429f5638b-kube-api-access-2cb9t\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.991805 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-bundle\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.991851 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-util\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.992891 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-bundle\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:55 crc kubenswrapper[4914]: I1122 14:38:55.992911 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-util\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:56 crc kubenswrapper[4914]: I1122 14:38:56.010493 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cb9t\" (UniqueName: \"kubernetes.io/projected/d0a62272-b520-435c-a59e-783429f5638b-kube-api-access-2cb9t\") pod \"7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:56 crc kubenswrapper[4914]: I1122 14:38:56.174820 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:38:56 crc kubenswrapper[4914]: I1122 14:38:56.619810 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl"] Nov 22 14:38:57 crc kubenswrapper[4914]: I1122 14:38:57.450629 4914 generic.go:334] "Generic (PLEG): container finished" podID="d0a62272-b520-435c-a59e-783429f5638b" containerID="bfd18e181608ee64a41d86346b00b6c653a57800a8d14ba66549f81bc4fefc35" exitCode=0 Nov 22 14:38:57 crc kubenswrapper[4914]: I1122 14:38:57.450683 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" event={"ID":"d0a62272-b520-435c-a59e-783429f5638b","Type":"ContainerDied","Data":"bfd18e181608ee64a41d86346b00b6c653a57800a8d14ba66549f81bc4fefc35"} Nov 22 14:38:57 crc kubenswrapper[4914]: I1122 14:38:57.451299 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" event={"ID":"d0a62272-b520-435c-a59e-783429f5638b","Type":"ContainerStarted","Data":"8d864f4461c7a27025765f8507714303d75e3e82173c6abc13eaeaef67cdd0a4"} Nov 22 14:38:58 crc kubenswrapper[4914]: I1122 14:38:58.459612 4914 generic.go:334] "Generic (PLEG): container finished" podID="d0a62272-b520-435c-a59e-783429f5638b" containerID="3159aa1a7df62a5a9beb51ee84ec46bd4e456ef4976e54f94708eba459040593" exitCode=0 Nov 22 14:38:58 crc kubenswrapper[4914]: I1122 14:38:58.459681 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" event={"ID":"d0a62272-b520-435c-a59e-783429f5638b","Type":"ContainerDied","Data":"3159aa1a7df62a5a9beb51ee84ec46bd4e456ef4976e54f94708eba459040593"} Nov 22 14:38:59 crc kubenswrapper[4914]: I1122 14:38:59.470028 4914 generic.go:334] "Generic (PLEG): container finished" podID="d0a62272-b520-435c-a59e-783429f5638b" containerID="e24c47a04c3af00e989225a4483aa1e6ba6b2a999cd5c81c6e408232800725ff" exitCode=0 Nov 22 14:38:59 crc kubenswrapper[4914]: I1122 14:38:59.470120 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" event={"ID":"d0a62272-b520-435c-a59e-783429f5638b","Type":"ContainerDied","Data":"e24c47a04c3af00e989225a4483aa1e6ba6b2a999cd5c81c6e408232800725ff"} Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.778996 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.860440 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cb9t\" (UniqueName: \"kubernetes.io/projected/d0a62272-b520-435c-a59e-783429f5638b-kube-api-access-2cb9t\") pod \"d0a62272-b520-435c-a59e-783429f5638b\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.860502 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-bundle\") pod \"d0a62272-b520-435c-a59e-783429f5638b\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.860562 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-util\") pod \"d0a62272-b520-435c-a59e-783429f5638b\" (UID: \"d0a62272-b520-435c-a59e-783429f5638b\") " Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.862229 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-bundle" (OuterVolumeSpecName: "bundle") pod "d0a62272-b520-435c-a59e-783429f5638b" (UID: "d0a62272-b520-435c-a59e-783429f5638b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.872096 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a62272-b520-435c-a59e-783429f5638b-kube-api-access-2cb9t" (OuterVolumeSpecName: "kube-api-access-2cb9t") pod "d0a62272-b520-435c-a59e-783429f5638b" (UID: "d0a62272-b520-435c-a59e-783429f5638b"). InnerVolumeSpecName "kube-api-access-2cb9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.885204 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-util" (OuterVolumeSpecName: "util") pod "d0a62272-b520-435c-a59e-783429f5638b" (UID: "d0a62272-b520-435c-a59e-783429f5638b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.962904 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.962975 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cb9t\" (UniqueName: \"kubernetes.io/projected/d0a62272-b520-435c-a59e-783429f5638b-kube-api-access-2cb9t\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:00 crc kubenswrapper[4914]: I1122 14:39:00.963004 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d0a62272-b520-435c-a59e-783429f5638b-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:01 crc kubenswrapper[4914]: I1122 14:39:01.486283 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" event={"ID":"d0a62272-b520-435c-a59e-783429f5638b","Type":"ContainerDied","Data":"8d864f4461c7a27025765f8507714303d75e3e82173c6abc13eaeaef67cdd0a4"} Nov 22 14:39:01 crc kubenswrapper[4914]: I1122 14:39:01.486340 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d864f4461c7a27025765f8507714303d75e3e82173c6abc13eaeaef67cdd0a4" Nov 22 14:39:01 crc kubenswrapper[4914]: I1122 14:39:01.486439 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.510829 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg"] Nov 22 14:39:04 crc kubenswrapper[4914]: E1122 14:39:04.511391 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="pull" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.511408 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="pull" Nov 22 14:39:04 crc kubenswrapper[4914]: E1122 14:39:04.511428 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="util" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.511434 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="util" Nov 22 14:39:04 crc kubenswrapper[4914]: E1122 14:39:04.511457 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="extract" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.511465 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="extract" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.511579 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a62272-b520-435c-a59e-783429f5638b" containerName="extract" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.512248 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.513974 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.514599 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.514823 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-h4rq6" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.531296 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg"] Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.609924 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10460829-745c-493b-938a-e7dbf0de739e-apiservice-cert\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.609985 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10460829-745c-493b-938a-e7dbf0de739e-webhook-cert\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.610125 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgrht\" (UniqueName: \"kubernetes.io/projected/10460829-745c-493b-938a-e7dbf0de739e-kube-api-access-zgrht\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.710676 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10460829-745c-493b-938a-e7dbf0de739e-apiservice-cert\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.710728 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10460829-745c-493b-938a-e7dbf0de739e-webhook-cert\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.710754 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgrht\" (UniqueName: \"kubernetes.io/projected/10460829-745c-493b-938a-e7dbf0de739e-kube-api-access-zgrht\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.715595 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10460829-745c-493b-938a-e7dbf0de739e-webhook-cert\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.720655 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10460829-745c-493b-938a-e7dbf0de739e-apiservice-cert\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.735273 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgrht\" (UniqueName: \"kubernetes.io/projected/10460829-745c-493b-938a-e7dbf0de739e-kube-api-access-zgrht\") pod \"mariadb-operator-controller-manager-7888767b5c-zp5lg\" (UID: \"10460829-745c-493b-938a-e7dbf0de739e\") " pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:04 crc kubenswrapper[4914]: I1122 14:39:04.830767 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:05 crc kubenswrapper[4914]: I1122 14:39:05.102272 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg"] Nov 22 14:39:05 crc kubenswrapper[4914]: I1122 14:39:05.510242 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" event={"ID":"10460829-745c-493b-938a-e7dbf0de739e","Type":"ContainerStarted","Data":"512896e2818906aeb8e5ca8cdcae8290f1abf5319412f2612e6cf706864baa3d"} Nov 22 14:39:09 crc kubenswrapper[4914]: I1122 14:39:09.533869 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" event={"ID":"10460829-745c-493b-938a-e7dbf0de739e","Type":"ContainerStarted","Data":"0e983621673f8408d5292318d61ed12db68af58fe363707b533faca0d3fb4a28"} Nov 22 14:39:11 crc kubenswrapper[4914]: I1122 14:39:11.545937 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" event={"ID":"10460829-745c-493b-938a-e7dbf0de739e","Type":"ContainerStarted","Data":"11cabdbd70255faa0bb4a72fcd3c0c33789c1cd50d7268af2bc54f4acb4c61fb"} Nov 22 14:39:11 crc kubenswrapper[4914]: I1122 14:39:11.546324 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:11 crc kubenswrapper[4914]: I1122 14:39:11.567771 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" podStartSLOduration=1.9556971010000002 podStartE2EDuration="7.56775572s" podCreationTimestamp="2025-11-22 14:39:04 +0000 UTC" firstStartedPulling="2025-11-22 14:39:05.111338839 +0000 UTC m=+710.919908500" lastFinishedPulling="2025-11-22 14:39:10.723397458 +0000 UTC m=+716.531967119" observedRunningTime="2025-11-22 14:39:11.562405141 +0000 UTC m=+717.370974802" watchObservedRunningTime="2025-11-22 14:39:11.56775572 +0000 UTC m=+717.376325381" Nov 22 14:39:14 crc kubenswrapper[4914]: I1122 14:39:14.836200 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7888767b5c-zp5lg" Nov 22 14:39:16 crc kubenswrapper[4914]: I1122 14:39:16.994015 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-2sm57"] Nov 22 14:39:16 crc kubenswrapper[4914]: I1122 14:39:16.995979 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.002712 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-2sm57"] Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.003420 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-fnsxr" Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.181985 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrxm\" (UniqueName: \"kubernetes.io/projected/eda159ed-53b5-4aeb-b122-087d6e823756-kube-api-access-5xrxm\") pod \"infra-operator-index-2sm57\" (UID: \"eda159ed-53b5-4aeb-b122-087d6e823756\") " pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.283444 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrxm\" (UniqueName: \"kubernetes.io/projected/eda159ed-53b5-4aeb-b122-087d6e823756-kube-api-access-5xrxm\") pod \"infra-operator-index-2sm57\" (UID: \"eda159ed-53b5-4aeb-b122-087d6e823756\") " pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.308579 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrxm\" (UniqueName: \"kubernetes.io/projected/eda159ed-53b5-4aeb-b122-087d6e823756-kube-api-access-5xrxm\") pod \"infra-operator-index-2sm57\" (UID: \"eda159ed-53b5-4aeb-b122-087d6e823756\") " pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.313157 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.563098 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-2sm57"] Nov 22 14:39:17 crc kubenswrapper[4914]: I1122 14:39:17.608114 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2sm57" event={"ID":"eda159ed-53b5-4aeb-b122-087d6e823756","Type":"ContainerStarted","Data":"6fcea7479063da9f14836b941466570c0c3c7465958c7e36ad07c982d0a9dd49"} Nov 22 14:39:19 crc kubenswrapper[4914]: I1122 14:39:19.621374 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2sm57" event={"ID":"eda159ed-53b5-4aeb-b122-087d6e823756","Type":"ContainerStarted","Data":"621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac"} Nov 22 14:39:19 crc kubenswrapper[4914]: I1122 14:39:19.638902 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-2sm57" podStartSLOduration=2.441093999 podStartE2EDuration="3.638856294s" podCreationTimestamp="2025-11-22 14:39:16 +0000 UTC" firstStartedPulling="2025-11-22 14:39:17.581489265 +0000 UTC m=+723.390058916" lastFinishedPulling="2025-11-22 14:39:18.77925152 +0000 UTC m=+724.587821211" observedRunningTime="2025-11-22 14:39:19.634425537 +0000 UTC m=+725.442995228" watchObservedRunningTime="2025-11-22 14:39:19.638856294 +0000 UTC m=+725.447425955" Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.178534 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-2sm57"] Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.784784 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-45zxz"] Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.785478 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.799759 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-45zxz"] Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.850072 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqrcz\" (UniqueName: \"kubernetes.io/projected/61fd0abd-4102-40bd-af29-8f800b1e4ef6-kube-api-access-tqrcz\") pod \"infra-operator-index-45zxz\" (UID: \"61fd0abd-4102-40bd-af29-8f800b1e4ef6\") " pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.951272 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqrcz\" (UniqueName: \"kubernetes.io/projected/61fd0abd-4102-40bd-af29-8f800b1e4ef6-kube-api-access-tqrcz\") pod \"infra-operator-index-45zxz\" (UID: \"61fd0abd-4102-40bd-af29-8f800b1e4ef6\") " pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:20 crc kubenswrapper[4914]: I1122 14:39:20.985085 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqrcz\" (UniqueName: \"kubernetes.io/projected/61fd0abd-4102-40bd-af29-8f800b1e4ef6-kube-api-access-tqrcz\") pod \"infra-operator-index-45zxz\" (UID: \"61fd0abd-4102-40bd-af29-8f800b1e4ef6\") " pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:21 crc kubenswrapper[4914]: I1122 14:39:21.108713 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:21 crc kubenswrapper[4914]: I1122 14:39:21.317418 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-45zxz"] Nov 22 14:39:21 crc kubenswrapper[4914]: W1122 14:39:21.329051 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61fd0abd_4102_40bd_af29_8f800b1e4ef6.slice/crio-dd1a80106f666e60bbaee7810c20fbc37d898a2959a0877977da0a35502d886a WatchSource:0}: Error finding container dd1a80106f666e60bbaee7810c20fbc37d898a2959a0877977da0a35502d886a: Status 404 returned error can't find the container with id dd1a80106f666e60bbaee7810c20fbc37d898a2959a0877977da0a35502d886a Nov 22 14:39:21 crc kubenswrapper[4914]: I1122 14:39:21.645010 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-45zxz" event={"ID":"61fd0abd-4102-40bd-af29-8f800b1e4ef6","Type":"ContainerStarted","Data":"dd1a80106f666e60bbaee7810c20fbc37d898a2959a0877977da0a35502d886a"} Nov 22 14:39:21 crc kubenswrapper[4914]: I1122 14:39:21.645183 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-2sm57" podUID="eda159ed-53b5-4aeb-b122-087d6e823756" containerName="registry-server" containerID="cri-o://621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac" gracePeriod=2 Nov 22 14:39:21 crc kubenswrapper[4914]: I1122 14:39:21.961483 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.042436 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.042497 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.063999 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xrxm\" (UniqueName: \"kubernetes.io/projected/eda159ed-53b5-4aeb-b122-087d6e823756-kube-api-access-5xrxm\") pod \"eda159ed-53b5-4aeb-b122-087d6e823756\" (UID: \"eda159ed-53b5-4aeb-b122-087d6e823756\") " Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.067700 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eda159ed-53b5-4aeb-b122-087d6e823756-kube-api-access-5xrxm" (OuterVolumeSpecName: "kube-api-access-5xrxm") pod "eda159ed-53b5-4aeb-b122-087d6e823756" (UID: "eda159ed-53b5-4aeb-b122-087d6e823756"). InnerVolumeSpecName "kube-api-access-5xrxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.164672 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xrxm\" (UniqueName: \"kubernetes.io/projected/eda159ed-53b5-4aeb-b122-087d6e823756-kube-api-access-5xrxm\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.652639 4914 generic.go:334] "Generic (PLEG): container finished" podID="eda159ed-53b5-4aeb-b122-087d6e823756" containerID="621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac" exitCode=0 Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.652686 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2sm57" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.652703 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2sm57" event={"ID":"eda159ed-53b5-4aeb-b122-087d6e823756","Type":"ContainerDied","Data":"621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac"} Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.652777 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2sm57" event={"ID":"eda159ed-53b5-4aeb-b122-087d6e823756","Type":"ContainerDied","Data":"6fcea7479063da9f14836b941466570c0c3c7465958c7e36ad07c982d0a9dd49"} Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.652809 4914 scope.go:117] "RemoveContainer" containerID="621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.655091 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-45zxz" event={"ID":"61fd0abd-4102-40bd-af29-8f800b1e4ef6","Type":"ContainerStarted","Data":"fc33977b5e1eeeb1ed3e3ebb94ff182b51b5b5db1dcc960b1d856c05ab2e757c"} Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.675469 4914 scope.go:117] "RemoveContainer" containerID="621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac" Nov 22 14:39:22 crc kubenswrapper[4914]: E1122 14:39:22.675913 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac\": container with ID starting with 621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac not found: ID does not exist" containerID="621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.675962 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac"} err="failed to get container status \"621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac\": rpc error: code = NotFound desc = could not find container \"621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac\": container with ID starting with 621a44b92787e7f9010e0f598649541e6eef847c80538f056f90bea12e16acac not found: ID does not exist" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.677669 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-45zxz" podStartSLOduration=2.24368617 podStartE2EDuration="2.677644013s" podCreationTimestamp="2025-11-22 14:39:20 +0000 UTC" firstStartedPulling="2025-11-22 14:39:21.332618504 +0000 UTC m=+727.141188185" lastFinishedPulling="2025-11-22 14:39:21.766576367 +0000 UTC m=+727.575146028" observedRunningTime="2025-11-22 14:39:22.672600371 +0000 UTC m=+728.481170032" watchObservedRunningTime="2025-11-22 14:39:22.677644013 +0000 UTC m=+728.486213704" Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.691330 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-2sm57"] Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.694285 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-2sm57"] Nov 22 14:39:22 crc kubenswrapper[4914]: I1122 14:39:22.873501 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eda159ed-53b5-4aeb-b122-087d6e823756" path="/var/lib/kubelet/pods/eda159ed-53b5-4aeb-b122-087d6e823756/volumes" Nov 22 14:39:31 crc kubenswrapper[4914]: I1122 14:39:31.109562 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:31 crc kubenswrapper[4914]: I1122 14:39:31.110186 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:31 crc kubenswrapper[4914]: I1122 14:39:31.136286 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:31 crc kubenswrapper[4914]: I1122 14:39:31.743512 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-45zxz" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.005320 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqw26"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.006163 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" podUID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" containerName="controller-manager" containerID="cri-o://d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b" gracePeriod=30 Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.102781 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.103041 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" podUID="6fc9cbad-c139-40ac-82c5-b076343c5288" containerName="route-controller-manager" containerID="cri-o://44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e" gracePeriod=30 Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.213591 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb"] Nov 22 14:39:39 crc kubenswrapper[4914]: E1122 14:39:39.213805 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda159ed-53b5-4aeb-b122-087d6e823756" containerName="registry-server" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.213816 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda159ed-53b5-4aeb-b122-087d6e823756" containerName="registry-server" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.213952 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda159ed-53b5-4aeb-b122-087d6e823756" containerName="registry-server" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.214667 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.218999 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l5bmb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.230994 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.381557 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-bundle\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.381933 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4rlc\" (UniqueName: \"kubernetes.io/projected/c1fb4fba-846d-443d-aa98-665954fd7733-kube-api-access-b4rlc\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.381977 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-util\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.423467 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.471893 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485164 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-client-ca\") pod \"6fc9cbad-c139-40ac-82c5-b076343c5288\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485228 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-config\") pod \"6fc9cbad-c139-40ac-82c5-b076343c5288\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485263 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-config\") pod \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485314 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttdtx\" (UniqueName: \"kubernetes.io/projected/6fc9cbad-c139-40ac-82c5-b076343c5288-kube-api-access-ttdtx\") pod \"6fc9cbad-c139-40ac-82c5-b076343c5288\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485349 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjwlf\" (UniqueName: \"kubernetes.io/projected/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-kube-api-access-qjwlf\") pod \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485378 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-client-ca\") pod \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485396 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9cbad-c139-40ac-82c5-b076343c5288-serving-cert\") pod \"6fc9cbad-c139-40ac-82c5-b076343c5288\" (UID: \"6fc9cbad-c139-40ac-82c5-b076343c5288\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485409 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-serving-cert\") pod \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485448 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-proxy-ca-bundles\") pod \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\" (UID: \"4d1efa71-4de9-43b1-8a6c-bcbf686b8057\") " Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485605 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-bundle\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485653 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4rlc\" (UniqueName: \"kubernetes.io/projected/c1fb4fba-846d-443d-aa98-665954fd7733-kube-api-access-b4rlc\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.485676 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-util\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.486097 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-client-ca" (OuterVolumeSpecName: "client-ca") pod "6fc9cbad-c139-40ac-82c5-b076343c5288" (UID: "6fc9cbad-c139-40ac-82c5-b076343c5288"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.486445 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-util\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.486596 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-client-ca" (OuterVolumeSpecName: "client-ca") pod "4d1efa71-4de9-43b1-8a6c-bcbf686b8057" (UID: "4d1efa71-4de9-43b1-8a6c-bcbf686b8057"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.487049 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-config" (OuterVolumeSpecName: "config") pod "6fc9cbad-c139-40ac-82c5-b076343c5288" (UID: "6fc9cbad-c139-40ac-82c5-b076343c5288"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.487506 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-config" (OuterVolumeSpecName: "config") pod "4d1efa71-4de9-43b1-8a6c-bcbf686b8057" (UID: "4d1efa71-4de9-43b1-8a6c-bcbf686b8057"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.488119 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-bundle\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.488686 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4d1efa71-4de9-43b1-8a6c-bcbf686b8057" (UID: "4d1efa71-4de9-43b1-8a6c-bcbf686b8057"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.493580 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fc9cbad-c139-40ac-82c5-b076343c5288-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6fc9cbad-c139-40ac-82c5-b076343c5288" (UID: "6fc9cbad-c139-40ac-82c5-b076343c5288"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.493739 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-kube-api-access-qjwlf" (OuterVolumeSpecName: "kube-api-access-qjwlf") pod "4d1efa71-4de9-43b1-8a6c-bcbf686b8057" (UID: "4d1efa71-4de9-43b1-8a6c-bcbf686b8057"). InnerVolumeSpecName "kube-api-access-qjwlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.502120 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4d1efa71-4de9-43b1-8a6c-bcbf686b8057" (UID: "4d1efa71-4de9-43b1-8a6c-bcbf686b8057"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.502388 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fc9cbad-c139-40ac-82c5-b076343c5288-kube-api-access-ttdtx" (OuterVolumeSpecName: "kube-api-access-ttdtx") pod "6fc9cbad-c139-40ac-82c5-b076343c5288" (UID: "6fc9cbad-c139-40ac-82c5-b076343c5288"). InnerVolumeSpecName "kube-api-access-ttdtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.526073 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4rlc\" (UniqueName: \"kubernetes.io/projected/c1fb4fba-846d-443d-aa98-665954fd7733-kube-api-access-b4rlc\") pod \"ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587385 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjwlf\" (UniqueName: \"kubernetes.io/projected/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-kube-api-access-qjwlf\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587433 4914 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587451 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9cbad-c139-40ac-82c5-b076343c5288-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587465 4914 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587477 4914 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587488 4914 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-client-ca\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587500 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fc9cbad-c139-40ac-82c5-b076343c5288-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587512 4914 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1efa71-4de9-43b1-8a6c-bcbf686b8057-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.587526 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttdtx\" (UniqueName: \"kubernetes.io/projected/6fc9cbad-c139-40ac-82c5-b076343c5288-kube-api-access-ttdtx\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.604853 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.785481 4914 generic.go:334] "Generic (PLEG): container finished" podID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" containerID="d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b" exitCode=0 Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.785556 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.785596 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" event={"ID":"4d1efa71-4de9-43b1-8a6c-bcbf686b8057","Type":"ContainerDied","Data":"d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b"} Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.785712 4914 scope.go:117] "RemoveContainer" containerID="d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.785765 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tqw26" event={"ID":"4d1efa71-4de9-43b1-8a6c-bcbf686b8057","Type":"ContainerDied","Data":"c57147b2d91d5186016e67f3e52a943d821206c6c34e5490107461cccdf60189"} Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.788146 4914 generic.go:334] "Generic (PLEG): container finished" podID="6fc9cbad-c139-40ac-82c5-b076343c5288" containerID="44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e" exitCode=0 Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.788215 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" event={"ID":"6fc9cbad-c139-40ac-82c5-b076343c5288","Type":"ContainerDied","Data":"44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e"} Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.788249 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" event={"ID":"6fc9cbad-c139-40ac-82c5-b076343c5288","Type":"ContainerDied","Data":"9c23132ba2155f034d73a62cbba5a44d371a0a3f9547be555ddd31fe14602a74"} Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.788325 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.805719 4914 scope.go:117] "RemoveContainer" containerID="d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b" Nov 22 14:39:39 crc kubenswrapper[4914]: E1122 14:39:39.807904 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b\": container with ID starting with d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b not found: ID does not exist" containerID="d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.807951 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b"} err="failed to get container status \"d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b\": rpc error: code = NotFound desc = could not find container \"d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b\": container with ID starting with d1e4cc26c35cd797504214a38ca8a6d182db5206c4b58b8f7940f5bd3f73c95b not found: ID does not exist" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.807978 4914 scope.go:117] "RemoveContainer" containerID="44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.824244 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqw26"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.828725 4914 scope.go:117] "RemoveContainer" containerID="44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e" Nov 22 14:39:39 crc kubenswrapper[4914]: E1122 14:39:39.829087 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e\": container with ID starting with 44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e not found: ID does not exist" containerID="44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.829132 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e"} err="failed to get container status \"44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e\": rpc error: code = NotFound desc = could not find container \"44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e\": container with ID starting with 44bd98ec9615c05f7e8a94c9c919f5b763c577253ddf1f62bcc7ed5f10c66f9e not found: ID does not exist" Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.829416 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tqw26"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.834460 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.836340 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr"] Nov 22 14:39:39 crc kubenswrapper[4914]: I1122 14:39:39.838865 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bpwxr"] Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.590596 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n"] Nov 22 14:39:40 crc kubenswrapper[4914]: E1122 14:39:40.591320 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" containerName="controller-manager" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.591339 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" containerName="controller-manager" Nov 22 14:39:40 crc kubenswrapper[4914]: E1122 14:39:40.591361 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fc9cbad-c139-40ac-82c5-b076343c5288" containerName="route-controller-manager" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.591370 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fc9cbad-c139-40ac-82c5-b076343c5288" containerName="route-controller-manager" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.591496 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fc9cbad-c139-40ac-82c5-b076343c5288" containerName="route-controller-manager" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.591522 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" containerName="controller-manager" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.591863 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.594109 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.595205 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.596362 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.596369 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.596364 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.596750 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.601778 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b168447e-6f9f-44f7-a0b1-ba3fd346c049-serving-cert\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.602038 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b168447e-6f9f-44f7-a0b1-ba3fd346c049-config\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.602183 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b168447e-6f9f-44f7-a0b1-ba3fd346c049-client-ca\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.602385 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd54p\" (UniqueName: \"kubernetes.io/projected/b168447e-6f9f-44f7-a0b1-ba3fd346c049-kube-api-access-fd54p\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.604132 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n"] Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.703209 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b168447e-6f9f-44f7-a0b1-ba3fd346c049-serving-cert\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.703251 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b168447e-6f9f-44f7-a0b1-ba3fd346c049-config\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.703278 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b168447e-6f9f-44f7-a0b1-ba3fd346c049-client-ca\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.703295 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd54p\" (UniqueName: \"kubernetes.io/projected/b168447e-6f9f-44f7-a0b1-ba3fd346c049-kube-api-access-fd54p\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.704380 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b168447e-6f9f-44f7-a0b1-ba3fd346c049-client-ca\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.704705 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b168447e-6f9f-44f7-a0b1-ba3fd346c049-config\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.712284 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b168447e-6f9f-44f7-a0b1-ba3fd346c049-serving-cert\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.718868 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd54p\" (UniqueName: \"kubernetes.io/projected/b168447e-6f9f-44f7-a0b1-ba3fd346c049-kube-api-access-fd54p\") pod \"route-controller-manager-6499fc8499-vr56n\" (UID: \"b168447e-6f9f-44f7-a0b1-ba3fd346c049\") " pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.794366 4914 generic.go:334] "Generic (PLEG): container finished" podID="c1fb4fba-846d-443d-aa98-665954fd7733" containerID="bf30d37cd2e9b546ee74163a7a125e60aef53c422ff8ad917e490d7e746f3304" exitCode=0 Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.794413 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" event={"ID":"c1fb4fba-846d-443d-aa98-665954fd7733","Type":"ContainerDied","Data":"bf30d37cd2e9b546ee74163a7a125e60aef53c422ff8ad917e490d7e746f3304"} Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.794461 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" event={"ID":"c1fb4fba-846d-443d-aa98-665954fd7733","Type":"ContainerStarted","Data":"2b67f1e7c864cd1bee02dbc4e8a4576c8c82f22990f7af3f6eed62e1945d25b6"} Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.875502 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d1efa71-4de9-43b1-8a6c-bcbf686b8057" path="/var/lib/kubelet/pods/4d1efa71-4de9-43b1-8a6c-bcbf686b8057/volumes" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.876366 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fc9cbad-c139-40ac-82c5-b076343c5288" path="/var/lib/kubelet/pods/6fc9cbad-c139-40ac-82c5-b076343c5288/volumes" Nov 22 14:39:40 crc kubenswrapper[4914]: I1122 14:39:40.915434 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.053470 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7685b8bc7f-zz98g"] Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.054410 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.058519 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.058679 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.058986 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.059068 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.059159 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.059323 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.090358 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.095614 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7685b8bc7f-zz98g"] Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.138561 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-proxy-ca-bundles\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.138617 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a68f96f-2852-4dae-be48-afdbcbb38525-serving-cert\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.138638 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-config\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.138662 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-client-ca\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.138683 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcmq4\" (UniqueName: \"kubernetes.io/projected/6a68f96f-2852-4dae-be48-afdbcbb38525-kube-api-access-vcmq4\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.212857 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n"] Nov 22 14:39:41 crc kubenswrapper[4914]: W1122 14:39:41.224974 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb168447e_6f9f_44f7_a0b1_ba3fd346c049.slice/crio-7e9014225a47554e96eaa66091c5efd58fd8a53e68010d55b066c44d581ec548 WatchSource:0}: Error finding container 7e9014225a47554e96eaa66091c5efd58fd8a53e68010d55b066c44d581ec548: Status 404 returned error can't find the container with id 7e9014225a47554e96eaa66091c5efd58fd8a53e68010d55b066c44d581ec548 Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.239528 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a68f96f-2852-4dae-be48-afdbcbb38525-serving-cert\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.239569 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-config\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.239596 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-client-ca\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.240094 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcmq4\" (UniqueName: \"kubernetes.io/projected/6a68f96f-2852-4dae-be48-afdbcbb38525-kube-api-access-vcmq4\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.240217 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-proxy-ca-bundles\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.240756 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-config\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.241297 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-client-ca\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.241450 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a68f96f-2852-4dae-be48-afdbcbb38525-proxy-ca-bundles\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.250332 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a68f96f-2852-4dae-be48-afdbcbb38525-serving-cert\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.263264 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcmq4\" (UniqueName: \"kubernetes.io/projected/6a68f96f-2852-4dae-be48-afdbcbb38525-kube-api-access-vcmq4\") pod \"controller-manager-7685b8bc7f-zz98g\" (UID: \"6a68f96f-2852-4dae-be48-afdbcbb38525\") " pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.384703 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.790060 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7685b8bc7f-zz98g"] Nov 22 14:39:41 crc kubenswrapper[4914]: W1122 14:39:41.805211 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a68f96f_2852_4dae_be48_afdbcbb38525.slice/crio-f5ee7115e0cbbaa4b88e586546a54ca7230fba927778d5b3814bd8f40943cb1e WatchSource:0}: Error finding container f5ee7115e0cbbaa4b88e586546a54ca7230fba927778d5b3814bd8f40943cb1e: Status 404 returned error can't find the container with id f5ee7115e0cbbaa4b88e586546a54ca7230fba927778d5b3814bd8f40943cb1e Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.806364 4914 generic.go:334] "Generic (PLEG): container finished" podID="c1fb4fba-846d-443d-aa98-665954fd7733" containerID="a9a8e70f0a8b6702fd0190c91397a09d8a6392a4735aa4938b1675fabdecccca" exitCode=0 Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.806514 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" event={"ID":"c1fb4fba-846d-443d-aa98-665954fd7733","Type":"ContainerDied","Data":"a9a8e70f0a8b6702fd0190c91397a09d8a6392a4735aa4938b1675fabdecccca"} Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.808191 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" event={"ID":"b168447e-6f9f-44f7-a0b1-ba3fd346c049","Type":"ContainerStarted","Data":"5888dac3e18d3f3ab3301b0985dc5c97eaa77452ed4105e61262bf74e5d47107"} Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.808223 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" event={"ID":"b168447e-6f9f-44f7-a0b1-ba3fd346c049","Type":"ContainerStarted","Data":"7e9014225a47554e96eaa66091c5efd58fd8a53e68010d55b066c44d581ec548"} Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.808440 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.813819 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" Nov 22 14:39:41 crc kubenswrapper[4914]: I1122 14:39:41.853565 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6499fc8499-vr56n" podStartSLOduration=1.8535469999999998 podStartE2EDuration="1.853547s" podCreationTimestamp="2025-11-22 14:39:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:39:41.852394441 +0000 UTC m=+747.660964132" watchObservedRunningTime="2025-11-22 14:39:41.853547 +0000 UTC m=+747.662116671" Nov 22 14:39:42 crc kubenswrapper[4914]: I1122 14:39:42.816542 4914 generic.go:334] "Generic (PLEG): container finished" podID="c1fb4fba-846d-443d-aa98-665954fd7733" containerID="daac74ed27aa4b3fadfd826382ef202c02d7ca9a8ec0102a4978999f85d9e76b" exitCode=0 Nov 22 14:39:42 crc kubenswrapper[4914]: I1122 14:39:42.816631 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" event={"ID":"c1fb4fba-846d-443d-aa98-665954fd7733","Type":"ContainerDied","Data":"daac74ed27aa4b3fadfd826382ef202c02d7ca9a8ec0102a4978999f85d9e76b"} Nov 22 14:39:42 crc kubenswrapper[4914]: I1122 14:39:42.818387 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" event={"ID":"6a68f96f-2852-4dae-be48-afdbcbb38525","Type":"ContainerStarted","Data":"7bc8d0ffc1b4545b0e5ac28a91b16f63f09dac2243286c3ecf7539bba1f3190d"} Nov 22 14:39:42 crc kubenswrapper[4914]: I1122 14:39:42.818424 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" event={"ID":"6a68f96f-2852-4dae-be48-afdbcbb38525","Type":"ContainerStarted","Data":"f5ee7115e0cbbaa4b88e586546a54ca7230fba927778d5b3814bd8f40943cb1e"} Nov 22 14:39:42 crc kubenswrapper[4914]: I1122 14:39:42.853007 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" podStartSLOduration=3.852986518 podStartE2EDuration="3.852986518s" podCreationTimestamp="2025-11-22 14:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:39:42.851964053 +0000 UTC m=+748.660533714" watchObservedRunningTime="2025-11-22 14:39:42.852986518 +0000 UTC m=+748.661556179" Nov 22 14:39:43 crc kubenswrapper[4914]: I1122 14:39:43.823862 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:43 crc kubenswrapper[4914]: I1122 14:39:43.831026 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7685b8bc7f-zz98g" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.122003 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.277347 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4rlc\" (UniqueName: \"kubernetes.io/projected/c1fb4fba-846d-443d-aa98-665954fd7733-kube-api-access-b4rlc\") pod \"c1fb4fba-846d-443d-aa98-665954fd7733\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.277411 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-util\") pod \"c1fb4fba-846d-443d-aa98-665954fd7733\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.277453 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-bundle\") pod \"c1fb4fba-846d-443d-aa98-665954fd7733\" (UID: \"c1fb4fba-846d-443d-aa98-665954fd7733\") " Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.278442 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-bundle" (OuterVolumeSpecName: "bundle") pod "c1fb4fba-846d-443d-aa98-665954fd7733" (UID: "c1fb4fba-846d-443d-aa98-665954fd7733"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.285398 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1fb4fba-846d-443d-aa98-665954fd7733-kube-api-access-b4rlc" (OuterVolumeSpecName: "kube-api-access-b4rlc") pod "c1fb4fba-846d-443d-aa98-665954fd7733" (UID: "c1fb4fba-846d-443d-aa98-665954fd7733"). InnerVolumeSpecName "kube-api-access-b4rlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.291595 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-util" (OuterVolumeSpecName: "util") pod "c1fb4fba-846d-443d-aa98-665954fd7733" (UID: "c1fb4fba-846d-443d-aa98-665954fd7733"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.378960 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4rlc\" (UniqueName: \"kubernetes.io/projected/c1fb4fba-846d-443d-aa98-665954fd7733-kube-api-access-b4rlc\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.379002 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.379016 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c1fb4fba-846d-443d-aa98-665954fd7733-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.841289 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" event={"ID":"c1fb4fba-846d-443d-aa98-665954fd7733","Type":"ContainerDied","Data":"2b67f1e7c864cd1bee02dbc4e8a4576c8c82f22990f7af3f6eed62e1945d25b6"} Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.841721 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b67f1e7c864cd1bee02dbc4e8a4576c8c82f22990f7af3f6eed62e1945d25b6" Nov 22 14:39:44 crc kubenswrapper[4914]: I1122 14:39:44.841307 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb" Nov 22 14:39:46 crc kubenswrapper[4914]: I1122 14:39:46.267992 4914 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.549768 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g"] Nov 22 14:39:51 crc kubenswrapper[4914]: E1122 14:39:51.552540 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="pull" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.552648 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="pull" Nov 22 14:39:51 crc kubenswrapper[4914]: E1122 14:39:51.552729 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="util" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.552828 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="util" Nov 22 14:39:51 crc kubenswrapper[4914]: E1122 14:39:51.552926 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="extract" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.553004 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="extract" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.553229 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1fb4fba-846d-443d-aa98-665954fd7733" containerName="extract" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.554280 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.560247 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-cclhn" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.562606 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.565907 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g"] Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.680591 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c72d9635-39e7-46df-9121-4f11c17f7ee5-apiservice-cert\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.680643 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwzj5\" (UniqueName: \"kubernetes.io/projected/c72d9635-39e7-46df-9121-4f11c17f7ee5-kube-api-access-hwzj5\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.680672 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c72d9635-39e7-46df-9121-4f11c17f7ee5-webhook-cert\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.782918 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwzj5\" (UniqueName: \"kubernetes.io/projected/c72d9635-39e7-46df-9121-4f11c17f7ee5-kube-api-access-hwzj5\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.783466 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c72d9635-39e7-46df-9121-4f11c17f7ee5-apiservice-cert\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.783665 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c72d9635-39e7-46df-9121-4f11c17f7ee5-webhook-cert\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.789129 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c72d9635-39e7-46df-9121-4f11c17f7ee5-apiservice-cert\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.789252 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c72d9635-39e7-46df-9121-4f11c17f7ee5-webhook-cert\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.805796 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwzj5\" (UniqueName: \"kubernetes.io/projected/c72d9635-39e7-46df-9121-4f11c17f7ee5-kube-api-access-hwzj5\") pod \"infra-operator-controller-manager-5b4b54c85c-k5v8g\" (UID: \"c72d9635-39e7-46df-9121-4f11c17f7ee5\") " pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:51 crc kubenswrapper[4914]: I1122 14:39:51.871542 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.043104 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.043202 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.043282 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.044516 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"448c460df7aa9c496e7a50b996cc4b57b27198da3e1c4c2cbdb4f3896aa006fa"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.044604 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://448c460df7aa9c496e7a50b996cc4b57b27198da3e1c4c2cbdb4f3896aa006fa" gracePeriod=600 Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.326594 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g"] Nov 22 14:39:52 crc kubenswrapper[4914]: W1122 14:39:52.336069 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc72d9635_39e7_46df_9121_4f11c17f7ee5.slice/crio-dec0f575186b7cd1eeea8966b41a33828a3ebf6e7663f24686cc141cd6446333 WatchSource:0}: Error finding container dec0f575186b7cd1eeea8966b41a33828a3ebf6e7663f24686cc141cd6446333: Status 404 returned error can't find the container with id dec0f575186b7cd1eeea8966b41a33828a3ebf6e7663f24686cc141cd6446333 Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.892043 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="448c460df7aa9c496e7a50b996cc4b57b27198da3e1c4c2cbdb4f3896aa006fa" exitCode=0 Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.892108 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"448c460df7aa9c496e7a50b996cc4b57b27198da3e1c4c2cbdb4f3896aa006fa"} Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.893139 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"aa8dd9653a1cc07eb97e4563bba7df9b08941e911b0fd3c2ed65154f5dab470f"} Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.893201 4914 scope.go:117] "RemoveContainer" containerID="8bb85fe0e7b9128d7e9a8dfe78ff109dee5730d63a76ca28de3d0e288cea71db" Nov 22 14:39:52 crc kubenswrapper[4914]: I1122 14:39:52.894928 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" event={"ID":"c72d9635-39e7-46df-9121-4f11c17f7ee5","Type":"ContainerStarted","Data":"dec0f575186b7cd1eeea8966b41a33828a3ebf6e7663f24686cc141cd6446333"} Nov 22 14:39:54 crc kubenswrapper[4914]: I1122 14:39:54.912198 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" event={"ID":"c72d9635-39e7-46df-9121-4f11c17f7ee5","Type":"ContainerStarted","Data":"bd5c963788e38a57f0834001b312e2b80918d28c6236ee1a6339c50cb07a2a67"} Nov 22 14:39:54 crc kubenswrapper[4914]: I1122 14:39:54.912620 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:39:54 crc kubenswrapper[4914]: I1122 14:39:54.912631 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" event={"ID":"c72d9635-39e7-46df-9121-4f11c17f7ee5","Type":"ContainerStarted","Data":"31efe59d9bcd34a6574ce65ee30361867b109f0302c4e0bf2dbbf5e7fba0806b"} Nov 22 14:39:54 crc kubenswrapper[4914]: I1122 14:39:54.934044 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" podStartSLOduration=2.49533017 podStartE2EDuration="3.934025819s" podCreationTimestamp="2025-11-22 14:39:51 +0000 UTC" firstStartedPulling="2025-11-22 14:39:52.338533342 +0000 UTC m=+758.147103003" lastFinishedPulling="2025-11-22 14:39:53.777228991 +0000 UTC m=+759.585798652" observedRunningTime="2025-11-22 14:39:54.932729678 +0000 UTC m=+760.741299329" watchObservedRunningTime="2025-11-22 14:39:54.934025819 +0000 UTC m=+760.742595480" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.212021 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.213942 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.219266 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-cbvxr" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.219419 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.219976 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.220104 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.221127 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.228663 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.231869 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.238122 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.239767 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.250188 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.262393 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.268521 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400524 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-kolla-config\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400578 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-operator-scripts\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400595 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-config-data-default\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400616 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sczx\" (UniqueName: \"kubernetes.io/projected/17cef1a9-a7c2-481e-82b0-6a6589c9462f-kube-api-access-6sczx\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400638 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a5d3cdb-0a84-422f-8195-59a5eaf47282-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400652 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm6zh\" (UniqueName: \"kubernetes.io/projected/1a2b9eb0-6a18-4786-b176-667b3147c524-kube-api-access-gm6zh\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400669 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/17cef1a9-a7c2-481e-82b0-6a6589c9462f-config-data-generated\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400689 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400707 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-kolla-config\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400721 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400738 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400757 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400774 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400796 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-config-data-default\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400812 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a2b9eb0-6a18-4786-b176-667b3147c524-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400828 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bmrt\" (UniqueName: \"kubernetes.io/projected/5a5d3cdb-0a84-422f-8195-59a5eaf47282-kube-api-access-8bmrt\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400844 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-config-data-default\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.400863 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-kolla-config\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.502571 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-kolla-config\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.502665 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-operator-scripts\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.502708 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-config-data-default\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.502760 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sczx\" (UniqueName: \"kubernetes.io/projected/17cef1a9-a7c2-481e-82b0-6a6589c9462f-kube-api-access-6sczx\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.502860 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a5d3cdb-0a84-422f-8195-59a5eaf47282-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.502953 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm6zh\" (UniqueName: \"kubernetes.io/projected/1a2b9eb0-6a18-4786-b176-667b3147c524-kube-api-access-gm6zh\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503038 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/17cef1a9-a7c2-481e-82b0-6a6589c9462f-config-data-generated\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503122 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503172 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-kolla-config\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503209 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503254 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503304 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503519 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/17cef1a9-a7c2-481e-82b0-6a6589c9462f-config-data-generated\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503605 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503530 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.503743 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5a5d3cdb-0a84-422f-8195-59a5eaf47282-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.504085 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-config-data-default\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.504141 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-kolla-config\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.504140 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.504331 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-kolla-config\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505209 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505218 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a2b9eb0-6a18-4786-b176-667b3147c524-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505220 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505328 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-config-data-default\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505368 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a2b9eb0-6a18-4786-b176-667b3147c524-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505403 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bmrt\" (UniqueName: \"kubernetes.io/projected/5a5d3cdb-0a84-422f-8195-59a5eaf47282-kube-api-access-8bmrt\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505444 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-config-data-default\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505493 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-kolla-config\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.505957 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a2b9eb0-6a18-4786-b176-667b3147c524-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.506134 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-kolla-config\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.507142 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-config-data-default\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.507465 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17cef1a9-a7c2-481e-82b0-6a6589c9462f-operator-scripts\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.508285 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5a5d3cdb-0a84-422f-8195-59a5eaf47282-config-data-default\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.529943 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.532247 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bmrt\" (UniqueName: \"kubernetes.io/projected/5a5d3cdb-0a84-422f-8195-59a5eaf47282-kube-api-access-8bmrt\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.534145 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm6zh\" (UniqueName: \"kubernetes.io/projected/1a2b9eb0-6a18-4786-b176-667b3147c524-kube-api-access-gm6zh\") pod \"openstack-galera-1\" (UID: \"1a2b9eb0-6a18-4786-b176-667b3147c524\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.536174 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sczx\" (UniqueName: \"kubernetes.io/projected/17cef1a9-a7c2-481e-82b0-6a6589c9462f-kube-api-access-6sczx\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.538011 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"5a5d3cdb-0a84-422f-8195-59a5eaf47282\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.542864 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"17cef1a9-a7c2-481e-82b0-6a6589c9462f\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.551636 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.562197 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:00 crc kubenswrapper[4914]: I1122 14:40:00.834795 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.029506 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 22 14:40:01 crc kubenswrapper[4914]: W1122 14:40:01.036383 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17cef1a9_a7c2_481e_82b0_6a6589c9462f.slice/crio-67da97a836f93b34e31fb3d35dcf62e42b31de374f2f73cf86606d8c53ab6963 WatchSource:0}: Error finding container 67da97a836f93b34e31fb3d35dcf62e42b31de374f2f73cf86606d8c53ab6963: Status 404 returned error can't find the container with id 67da97a836f93b34e31fb3d35dcf62e42b31de374f2f73cf86606d8c53ab6963 Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.114963 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 22 14:40:01 crc kubenswrapper[4914]: W1122 14:40:01.120859 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2b9eb0_6a18_4786_b176_667b3147c524.slice/crio-58feaa67361135e266cb545ad4c1fa852e7ffb738d9a39b5f545b8df7c15eb87 WatchSource:0}: Error finding container 58feaa67361135e266cb545ad4c1fa852e7ffb738d9a39b5f545b8df7c15eb87: Status 404 returned error can't find the container with id 58feaa67361135e266cb545ad4c1fa852e7ffb738d9a39b5f545b8df7c15eb87 Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.280557 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 22 14:40:01 crc kubenswrapper[4914]: W1122 14:40:01.285693 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a5d3cdb_0a84_422f_8195_59a5eaf47282.slice/crio-0513fc1cc44055471b6c3b45a7de74eb599fa3528ef10927a53c571824aa2423 WatchSource:0}: Error finding container 0513fc1cc44055471b6c3b45a7de74eb599fa3528ef10927a53c571824aa2423: Status 404 returned error can't find the container with id 0513fc1cc44055471b6c3b45a7de74eb599fa3528ef10927a53c571824aa2423 Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.877289 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5b4b54c85c-k5v8g" Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.967666 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"17cef1a9-a7c2-481e-82b0-6a6589c9462f","Type":"ContainerStarted","Data":"67da97a836f93b34e31fb3d35dcf62e42b31de374f2f73cf86606d8c53ab6963"} Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.968932 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5a5d3cdb-0a84-422f-8195-59a5eaf47282","Type":"ContainerStarted","Data":"0513fc1cc44055471b6c3b45a7de74eb599fa3528ef10927a53c571824aa2423"} Nov 22 14:40:01 crc kubenswrapper[4914]: I1122 14:40:01.970202 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1a2b9eb0-6a18-4786-b176-667b3147c524","Type":"ContainerStarted","Data":"58feaa67361135e266cb545ad4c1fa852e7ffb738d9a39b5f545b8df7c15eb87"} Nov 22 14:40:06 crc kubenswrapper[4914]: I1122 14:40:06.967275 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 22 14:40:06 crc kubenswrapper[4914]: I1122 14:40:06.968377 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:06 crc kubenswrapper[4914]: I1122 14:40:06.975420 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 22 14:40:06 crc kubenswrapper[4914]: I1122 14:40:06.975702 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-ndlqx" Nov 22 14:40:06 crc kubenswrapper[4914]: I1122 14:40:06.983442 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.010137 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9h6f\" (UniqueName: \"kubernetes.io/projected/c67b2123-2f5f-484e-9db5-3b74583d4047-kube-api-access-h9h6f\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.010205 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c67b2123-2f5f-484e-9db5-3b74583d4047-kolla-config\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.010223 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c67b2123-2f5f-484e-9db5-3b74583d4047-config-data\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.111706 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9h6f\" (UniqueName: \"kubernetes.io/projected/c67b2123-2f5f-484e-9db5-3b74583d4047-kube-api-access-h9h6f\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.111778 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c67b2123-2f5f-484e-9db5-3b74583d4047-kolla-config\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.111800 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c67b2123-2f5f-484e-9db5-3b74583d4047-config-data\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.112699 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c67b2123-2f5f-484e-9db5-3b74583d4047-kolla-config\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.112783 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c67b2123-2f5f-484e-9db5-3b74583d4047-config-data\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.144926 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9h6f\" (UniqueName: \"kubernetes.io/projected/c67b2123-2f5f-484e-9db5-3b74583d4047-kube-api-access-h9h6f\") pod \"memcached-0\" (UID: \"c67b2123-2f5f-484e-9db5-3b74583d4047\") " pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.283686 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.904739 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-mb5s5"] Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.905443 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.907179 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-z5n62" Nov 22 14:40:07 crc kubenswrapper[4914]: I1122 14:40:07.913681 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-mb5s5"] Nov 22 14:40:08 crc kubenswrapper[4914]: I1122 14:40:08.021415 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msk9w\" (UniqueName: \"kubernetes.io/projected/d0e6162b-5221-4a6a-ad3b-23c174b521e7-kube-api-access-msk9w\") pod \"rabbitmq-cluster-operator-index-mb5s5\" (UID: \"d0e6162b-5221-4a6a-ad3b-23c174b521e7\") " pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:08 crc kubenswrapper[4914]: I1122 14:40:08.123317 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msk9w\" (UniqueName: \"kubernetes.io/projected/d0e6162b-5221-4a6a-ad3b-23c174b521e7-kube-api-access-msk9w\") pod \"rabbitmq-cluster-operator-index-mb5s5\" (UID: \"d0e6162b-5221-4a6a-ad3b-23c174b521e7\") " pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:08 crc kubenswrapper[4914]: I1122 14:40:08.140086 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msk9w\" (UniqueName: \"kubernetes.io/projected/d0e6162b-5221-4a6a-ad3b-23c174b521e7-kube-api-access-msk9w\") pod \"rabbitmq-cluster-operator-index-mb5s5\" (UID: \"d0e6162b-5221-4a6a-ad3b-23c174b521e7\") " pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:08 crc kubenswrapper[4914]: I1122 14:40:08.222342 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:09 crc kubenswrapper[4914]: I1122 14:40:09.456458 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-mb5s5"] Nov 22 14:40:09 crc kubenswrapper[4914]: W1122 14:40:09.467008 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0e6162b_5221_4a6a_ad3b_23c174b521e7.slice/crio-c0c4c0fba1bdde3d60937bf45d0174e8860be9dcb1bac5ea7f318b8eaec142d6 WatchSource:0}: Error finding container c0c4c0fba1bdde3d60937bf45d0174e8860be9dcb1bac5ea7f318b8eaec142d6: Status 404 returned error can't find the container with id c0c4c0fba1bdde3d60937bf45d0174e8860be9dcb1bac5ea7f318b8eaec142d6 Nov 22 14:40:09 crc kubenswrapper[4914]: I1122 14:40:09.540598 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 22 14:40:10 crc kubenswrapper[4914]: I1122 14:40:10.029039 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1a2b9eb0-6a18-4786-b176-667b3147c524","Type":"ContainerStarted","Data":"8963d41844e3b6ccd6d5de24adcac8fc0fa0dde57d73638c763b25215d84ae9c"} Nov 22 14:40:10 crc kubenswrapper[4914]: I1122 14:40:10.030534 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" event={"ID":"d0e6162b-5221-4a6a-ad3b-23c174b521e7","Type":"ContainerStarted","Data":"c0c4c0fba1bdde3d60937bf45d0174e8860be9dcb1bac5ea7f318b8eaec142d6"} Nov 22 14:40:10 crc kubenswrapper[4914]: I1122 14:40:10.031565 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"c67b2123-2f5f-484e-9db5-3b74583d4047","Type":"ContainerStarted","Data":"55a8bf368688b9bf14a01448e78b6af3fa35edbc854c8a2a72a3c631881744f9"} Nov 22 14:40:10 crc kubenswrapper[4914]: I1122 14:40:10.033284 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"17cef1a9-a7c2-481e-82b0-6a6589c9462f","Type":"ContainerStarted","Data":"a6439bd3ceaab29b5b0f55df3fd927fbc577b00b19ae72cf17ecdf619c39db4d"} Nov 22 14:40:10 crc kubenswrapper[4914]: I1122 14:40:10.034305 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5a5d3cdb-0a84-422f-8195-59a5eaf47282","Type":"ContainerStarted","Data":"43b26f37712e6e60e41eb97331b893f41860b8b9715011ab99c36c9ed345ed09"} Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.100412 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-mb5s5"] Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.705380 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wn6fr"] Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.706371 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.748547 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wn6fr"] Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.792676 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhp9c\" (UniqueName: \"kubernetes.io/projected/37ed417d-6cea-4336-9cad-218ec6fa8b12-kube-api-access-dhp9c\") pod \"rabbitmq-cluster-operator-index-wn6fr\" (UID: \"37ed417d-6cea-4336-9cad-218ec6fa8b12\") " pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.894055 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhp9c\" (UniqueName: \"kubernetes.io/projected/37ed417d-6cea-4336-9cad-218ec6fa8b12-kube-api-access-dhp9c\") pod \"rabbitmq-cluster-operator-index-wn6fr\" (UID: \"37ed417d-6cea-4336-9cad-218ec6fa8b12\") " pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:12 crc kubenswrapper[4914]: I1122 14:40:12.912647 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhp9c\" (UniqueName: \"kubernetes.io/projected/37ed417d-6cea-4336-9cad-218ec6fa8b12-kube-api-access-dhp9c\") pod \"rabbitmq-cluster-operator-index-wn6fr\" (UID: \"37ed417d-6cea-4336-9cad-218ec6fa8b12\") " pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.035955 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.068023 4914 generic.go:334] "Generic (PLEG): container finished" podID="17cef1a9-a7c2-481e-82b0-6a6589c9462f" containerID="a6439bd3ceaab29b5b0f55df3fd927fbc577b00b19ae72cf17ecdf619c39db4d" exitCode=0 Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.068073 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"17cef1a9-a7c2-481e-82b0-6a6589c9462f","Type":"ContainerDied","Data":"a6439bd3ceaab29b5b0f55df3fd927fbc577b00b19ae72cf17ecdf619c39db4d"} Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.070445 4914 generic.go:334] "Generic (PLEG): container finished" podID="5a5d3cdb-0a84-422f-8195-59a5eaf47282" containerID="43b26f37712e6e60e41eb97331b893f41860b8b9715011ab99c36c9ed345ed09" exitCode=0 Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.070648 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5a5d3cdb-0a84-422f-8195-59a5eaf47282","Type":"ContainerDied","Data":"43b26f37712e6e60e41eb97331b893f41860b8b9715011ab99c36c9ed345ed09"} Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.072852 4914 generic.go:334] "Generic (PLEG): container finished" podID="1a2b9eb0-6a18-4786-b176-667b3147c524" containerID="8963d41844e3b6ccd6d5de24adcac8fc0fa0dde57d73638c763b25215d84ae9c" exitCode=0 Nov 22 14:40:13 crc kubenswrapper[4914]: I1122 14:40:13.072896 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1a2b9eb0-6a18-4786-b176-667b3147c524","Type":"ContainerDied","Data":"8963d41844e3b6ccd6d5de24adcac8fc0fa0dde57d73638c763b25215d84ae9c"} Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.108833 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" event={"ID":"d0e6162b-5221-4a6a-ad3b-23c174b521e7","Type":"ContainerStarted","Data":"38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9"} Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.108925 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" podUID="d0e6162b-5221-4a6a-ad3b-23c174b521e7" containerName="registry-server" containerID="cri-o://38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9" gracePeriod=2 Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.111041 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"c67b2123-2f5f-484e-9db5-3b74583d4047","Type":"ContainerStarted","Data":"8b4b9dd8ed1a176cf264ecaccac43a47a0038d4f4b8990f3ef5fbb0146d7d927"} Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.111141 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.114615 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"17cef1a9-a7c2-481e-82b0-6a6589c9462f","Type":"ContainerStarted","Data":"007f25db0dde5b88744437afed5065c88f32392c7d582b8a775beb1321e5b981"} Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.119393 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"5a5d3cdb-0a84-422f-8195-59a5eaf47282","Type":"ContainerStarted","Data":"40056e03ad45f7bb04ecf9f4d9b20fb72025733e7f885dd5d7839a0d6840a6cb"} Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.120966 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"1a2b9eb0-6a18-4786-b176-667b3147c524","Type":"ContainerStarted","Data":"0be5036793273362ce6e11f9f541d8a893077f5a6e266c6a2775ea38c98df326"} Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.130976 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" podStartSLOduration=2.850143817 podStartE2EDuration="12.130960059s" podCreationTimestamp="2025-11-22 14:40:07 +0000 UTC" firstStartedPulling="2025-11-22 14:40:09.468559643 +0000 UTC m=+775.277129304" lastFinishedPulling="2025-11-22 14:40:18.749375885 +0000 UTC m=+784.557945546" observedRunningTime="2025-11-22 14:40:19.129866942 +0000 UTC m=+784.938436603" watchObservedRunningTime="2025-11-22 14:40:19.130960059 +0000 UTC m=+784.939529730" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.152178 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=12.095191356 podStartE2EDuration="20.152164387s" podCreationTimestamp="2025-11-22 14:39:59 +0000 UTC" firstStartedPulling="2025-11-22 14:40:01.038517802 +0000 UTC m=+766.847087463" lastFinishedPulling="2025-11-22 14:40:09.095490793 +0000 UTC m=+774.904060494" observedRunningTime="2025-11-22 14:40:19.150559448 +0000 UTC m=+784.959129109" watchObservedRunningTime="2025-11-22 14:40:19.152164387 +0000 UTC m=+784.960734048" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.174740 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=3.879803416 podStartE2EDuration="13.174705187s" podCreationTimestamp="2025-11-22 14:40:06 +0000 UTC" firstStartedPulling="2025-11-22 14:40:09.540056084 +0000 UTC m=+775.348625745" lastFinishedPulling="2025-11-22 14:40:18.834957855 +0000 UTC m=+784.643527516" observedRunningTime="2025-11-22 14:40:19.172975195 +0000 UTC m=+784.981544866" watchObservedRunningTime="2025-11-22 14:40:19.174705187 +0000 UTC m=+784.983274848" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.191484 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wn6fr"] Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.193978 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=12.34747126 podStartE2EDuration="20.193962047s" podCreationTimestamp="2025-11-22 14:39:59 +0000 UTC" firstStartedPulling="2025-11-22 14:40:01.28820564 +0000 UTC m=+767.096775301" lastFinishedPulling="2025-11-22 14:40:09.134696417 +0000 UTC m=+774.943266088" observedRunningTime="2025-11-22 14:40:19.19285203 +0000 UTC m=+785.001421691" watchObservedRunningTime="2025-11-22 14:40:19.193962047 +0000 UTC m=+785.002531708" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.215118 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=12.265316922 podStartE2EDuration="20.215104013s" podCreationTimestamp="2025-11-22 14:39:59 +0000 UTC" firstStartedPulling="2025-11-22 14:40:01.123407869 +0000 UTC m=+766.931977530" lastFinishedPulling="2025-11-22 14:40:09.07319496 +0000 UTC m=+774.881764621" observedRunningTime="2025-11-22 14:40:19.210656375 +0000 UTC m=+785.019226036" watchObservedRunningTime="2025-11-22 14:40:19.215104013 +0000 UTC m=+785.023673674" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.473720 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.596694 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msk9w\" (UniqueName: \"kubernetes.io/projected/d0e6162b-5221-4a6a-ad3b-23c174b521e7-kube-api-access-msk9w\") pod \"d0e6162b-5221-4a6a-ad3b-23c174b521e7\" (UID: \"d0e6162b-5221-4a6a-ad3b-23c174b521e7\") " Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.601506 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0e6162b-5221-4a6a-ad3b-23c174b521e7-kube-api-access-msk9w" (OuterVolumeSpecName: "kube-api-access-msk9w") pod "d0e6162b-5221-4a6a-ad3b-23c174b521e7" (UID: "d0e6162b-5221-4a6a-ad3b-23c174b521e7"). InnerVolumeSpecName "kube-api-access-msk9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:40:19 crc kubenswrapper[4914]: I1122 14:40:19.698816 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msk9w\" (UniqueName: \"kubernetes.io/projected/d0e6162b-5221-4a6a-ad3b-23c174b521e7-kube-api-access-msk9w\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.128476 4914 generic.go:334] "Generic (PLEG): container finished" podID="d0e6162b-5221-4a6a-ad3b-23c174b521e7" containerID="38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9" exitCode=0 Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.128579 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.128953 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" event={"ID":"d0e6162b-5221-4a6a-ad3b-23c174b521e7","Type":"ContainerDied","Data":"38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9"} Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.129031 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-mb5s5" event={"ID":"d0e6162b-5221-4a6a-ad3b-23c174b521e7","Type":"ContainerDied","Data":"c0c4c0fba1bdde3d60937bf45d0174e8860be9dcb1bac5ea7f318b8eaec142d6"} Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.129065 4914 scope.go:117] "RemoveContainer" containerID="38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.131164 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" event={"ID":"37ed417d-6cea-4336-9cad-218ec6fa8b12","Type":"ContainerStarted","Data":"40a6fbe15f794e3b7ecb65069a51411ff2aa2c4f7839ff8eced93a72878374d0"} Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.131204 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" event={"ID":"37ed417d-6cea-4336-9cad-218ec6fa8b12","Type":"ContainerStarted","Data":"8af19ade8f7baacad5e6837ba161052693813a47c8af45ba15b961facb00ab61"} Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.159470 4914 scope.go:117] "RemoveContainer" containerID="38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9" Nov 22 14:40:20 crc kubenswrapper[4914]: E1122 14:40:20.162521 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9\": container with ID starting with 38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9 not found: ID does not exist" containerID="38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.162706 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9"} err="failed to get container status \"38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9\": rpc error: code = NotFound desc = could not find container \"38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9\": container with ID starting with 38409243deb5c61d4350985342185b1aaaba475c58bbae0c441203a372099da9 not found: ID does not exist" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.169714 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" podStartSLOduration=7.722810354 podStartE2EDuration="8.169689372s" podCreationTimestamp="2025-11-22 14:40:12 +0000 UTC" firstStartedPulling="2025-11-22 14:40:19.201559382 +0000 UTC m=+785.010129043" lastFinishedPulling="2025-11-22 14:40:19.64843836 +0000 UTC m=+785.457008061" observedRunningTime="2025-11-22 14:40:20.160016706 +0000 UTC m=+785.968586457" watchObservedRunningTime="2025-11-22 14:40:20.169689372 +0000 UTC m=+785.978259033" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.190622 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-mb5s5"] Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.195288 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-mb5s5"] Nov 22 14:40:20 crc kubenswrapper[4914]: E1122 14:40:20.397319 4914 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.98:44020->38.102.83.98:44331: write tcp 38.102.83.98:44020->38.102.83.98:44331: write: broken pipe Nov 22 14:40:20 crc kubenswrapper[4914]: E1122 14:40:20.490894 4914 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.98:44040->38.102.83.98:44331: write tcp 38.102.83.98:44040->38.102.83.98:44331: write: broken pipe Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.552546 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.552754 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.563414 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.563847 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.836115 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.836191 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:20 crc kubenswrapper[4914]: I1122 14:40:20.876297 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0e6162b-5221-4a6a-ad3b-23c174b521e7" path="/var/lib/kubelet/pods/d0e6162b-5221-4a6a-ad3b-23c174b521e7/volumes" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.121152 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wc2d4"] Nov 22 14:40:21 crc kubenswrapper[4914]: E1122 14:40:21.121744 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e6162b-5221-4a6a-ad3b-23c174b521e7" containerName="registry-server" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.121808 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e6162b-5221-4a6a-ad3b-23c174b521e7" containerName="registry-server" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.121979 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0e6162b-5221-4a6a-ad3b-23c174b521e7" containerName="registry-server" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.122818 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.133994 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc2d4"] Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.219008 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-catalog-content\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.219072 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-utilities\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.219159 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqh65\" (UniqueName: \"kubernetes.io/projected/e7f0d439-2c8c-42d1-8cdc-84df2101407f-kube-api-access-kqh65\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.320496 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqh65\" (UniqueName: \"kubernetes.io/projected/e7f0d439-2c8c-42d1-8cdc-84df2101407f-kube-api-access-kqh65\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.320599 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-catalog-content\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.320633 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-utilities\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.321139 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-utilities\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.321245 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-catalog-content\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.340994 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqh65\" (UniqueName: \"kubernetes.io/projected/e7f0d439-2c8c-42d1-8cdc-84df2101407f-kube-api-access-kqh65\") pod \"community-operators-wc2d4\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.443571 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:21 crc kubenswrapper[4914]: I1122 14:40:21.889257 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc2d4"] Nov 22 14:40:21 crc kubenswrapper[4914]: W1122 14:40:21.897052 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7f0d439_2c8c_42d1_8cdc_84df2101407f.slice/crio-c8154d9a5d590b748bd39ab442e04367c71f0b5090bfaf83a65e0da52fcb9a0e WatchSource:0}: Error finding container c8154d9a5d590b748bd39ab442e04367c71f0b5090bfaf83a65e0da52fcb9a0e: Status 404 returned error can't find the container with id c8154d9a5d590b748bd39ab442e04367c71f0b5090bfaf83a65e0da52fcb9a0e Nov 22 14:40:22 crc kubenswrapper[4914]: I1122 14:40:22.152480 4914 generic.go:334] "Generic (PLEG): container finished" podID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerID="b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5" exitCode=0 Nov 22 14:40:22 crc kubenswrapper[4914]: I1122 14:40:22.152544 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerDied","Data":"b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5"} Nov 22 14:40:22 crc kubenswrapper[4914]: I1122 14:40:22.152755 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerStarted","Data":"c8154d9a5d590b748bd39ab442e04367c71f0b5090bfaf83a65e0da52fcb9a0e"} Nov 22 14:40:23 crc kubenswrapper[4914]: I1122 14:40:23.036254 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:23 crc kubenswrapper[4914]: I1122 14:40:23.036302 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:23 crc kubenswrapper[4914]: I1122 14:40:23.070404 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:23 crc kubenswrapper[4914]: I1122 14:40:23.159633 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerStarted","Data":"182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4"} Nov 22 14:40:24 crc kubenswrapper[4914]: I1122 14:40:24.167822 4914 generic.go:334] "Generic (PLEG): container finished" podID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerID="182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4" exitCode=0 Nov 22 14:40:24 crc kubenswrapper[4914]: I1122 14:40:24.167933 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerDied","Data":"182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4"} Nov 22 14:40:25 crc kubenswrapper[4914]: I1122 14:40:25.176166 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerStarted","Data":"f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048"} Nov 22 14:40:25 crc kubenswrapper[4914]: I1122 14:40:25.199726 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wc2d4" podStartSLOduration=1.630328321 podStartE2EDuration="4.199707192s" podCreationTimestamp="2025-11-22 14:40:21 +0000 UTC" firstStartedPulling="2025-11-22 14:40:22.154062186 +0000 UTC m=+787.962631847" lastFinishedPulling="2025-11-22 14:40:24.723441057 +0000 UTC m=+790.532010718" observedRunningTime="2025-11-22 14:40:25.196199077 +0000 UTC m=+791.004768748" watchObservedRunningTime="2025-11-22 14:40:25.199707192 +0000 UTC m=+791.008276863" Nov 22 14:40:26 crc kubenswrapper[4914]: I1122 14:40:26.650102 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:26 crc kubenswrapper[4914]: I1122 14:40:26.763583 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 22 14:40:27 crc kubenswrapper[4914]: I1122 14:40:27.285243 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 22 14:40:31 crc kubenswrapper[4914]: I1122 14:40:31.443751 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:31 crc kubenswrapper[4914]: I1122 14:40:31.444025 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:31 crc kubenswrapper[4914]: I1122 14:40:31.479285 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:32 crc kubenswrapper[4914]: I1122 14:40:32.273424 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:33 crc kubenswrapper[4914]: I1122 14:40:33.060992 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-wn6fr" Nov 22 14:40:35 crc kubenswrapper[4914]: I1122 14:40:35.501105 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wc2d4"] Nov 22 14:40:35 crc kubenswrapper[4914]: I1122 14:40:35.501378 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wc2d4" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="registry-server" containerID="cri-o://f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048" gracePeriod=2 Nov 22 14:40:35 crc kubenswrapper[4914]: I1122 14:40:35.921064 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.026462 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-catalog-content\") pod \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.026810 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-utilities\") pod \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.027094 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqh65\" (UniqueName: \"kubernetes.io/projected/e7f0d439-2c8c-42d1-8cdc-84df2101407f-kube-api-access-kqh65\") pod \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\" (UID: \"e7f0d439-2c8c-42d1-8cdc-84df2101407f\") " Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.028315 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-utilities" (OuterVolumeSpecName: "utilities") pod "e7f0d439-2c8c-42d1-8cdc-84df2101407f" (UID: "e7f0d439-2c8c-42d1-8cdc-84df2101407f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.034673 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f0d439-2c8c-42d1-8cdc-84df2101407f-kube-api-access-kqh65" (OuterVolumeSpecName: "kube-api-access-kqh65") pod "e7f0d439-2c8c-42d1-8cdc-84df2101407f" (UID: "e7f0d439-2c8c-42d1-8cdc-84df2101407f"). InnerVolumeSpecName "kube-api-access-kqh65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.074536 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7f0d439-2c8c-42d1-8cdc-84df2101407f" (UID: "e7f0d439-2c8c-42d1-8cdc-84df2101407f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.128767 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqh65\" (UniqueName: \"kubernetes.io/projected/e7f0d439-2c8c-42d1-8cdc-84df2101407f-kube-api-access-kqh65\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.128808 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.128823 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f0d439-2c8c-42d1-8cdc-84df2101407f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.255045 4914 generic.go:334] "Generic (PLEG): container finished" podID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerID="f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048" exitCode=0 Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.255096 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc2d4" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.255101 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerDied","Data":"f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048"} Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.255229 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc2d4" event={"ID":"e7f0d439-2c8c-42d1-8cdc-84df2101407f","Type":"ContainerDied","Data":"c8154d9a5d590b748bd39ab442e04367c71f0b5090bfaf83a65e0da52fcb9a0e"} Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.255259 4914 scope.go:117] "RemoveContainer" containerID="f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.280728 4914 scope.go:117] "RemoveContainer" containerID="182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.287100 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wc2d4"] Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.292306 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wc2d4"] Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.319453 4914 scope.go:117] "RemoveContainer" containerID="b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.336497 4914 scope.go:117] "RemoveContainer" containerID="f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048" Nov 22 14:40:36 crc kubenswrapper[4914]: E1122 14:40:36.336832 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048\": container with ID starting with f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048 not found: ID does not exist" containerID="f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.336860 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048"} err="failed to get container status \"f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048\": rpc error: code = NotFound desc = could not find container \"f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048\": container with ID starting with f33bda647f9309072683ef0d2b4d3859822cb368f3c8c0535b37842dc22db048 not found: ID does not exist" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.336893 4914 scope.go:117] "RemoveContainer" containerID="182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4" Nov 22 14:40:36 crc kubenswrapper[4914]: E1122 14:40:36.337073 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4\": container with ID starting with 182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4 not found: ID does not exist" containerID="182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.337099 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4"} err="failed to get container status \"182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4\": rpc error: code = NotFound desc = could not find container \"182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4\": container with ID starting with 182c7b8ce74eb9a94fb9cf3be0875ff90d2e723cccf30a79f1193a58876532e4 not found: ID does not exist" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.337113 4914 scope.go:117] "RemoveContainer" containerID="b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5" Nov 22 14:40:36 crc kubenswrapper[4914]: E1122 14:40:36.337286 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5\": container with ID starting with b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5 not found: ID does not exist" containerID="b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.337306 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5"} err="failed to get container status \"b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5\": rpc error: code = NotFound desc = could not find container \"b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5\": container with ID starting with b0baae8eff60bc2caa016b09aabef3be325067157fd13c870b2787515e5557d5 not found: ID does not exist" Nov 22 14:40:36 crc kubenswrapper[4914]: I1122 14:40:36.881829 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" path="/var/lib/kubelet/pods/e7f0d439-2c8c-42d1-8cdc-84df2101407f/volumes" Nov 22 14:40:40 crc kubenswrapper[4914]: I1122 14:40:40.087414 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:40 crc kubenswrapper[4914]: I1122 14:40:40.166572 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 22 14:40:41 crc kubenswrapper[4914]: I1122 14:40:41.397156 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:41 crc kubenswrapper[4914]: I1122 14:40:41.474690 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.552682 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh"] Nov 22 14:40:43 crc kubenswrapper[4914]: E1122 14:40:43.553386 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="extract-content" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.553403 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="extract-content" Nov 22 14:40:43 crc kubenswrapper[4914]: E1122 14:40:43.553421 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="registry-server" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.553429 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="registry-server" Nov 22 14:40:43 crc kubenswrapper[4914]: E1122 14:40:43.553440 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="extract-utilities" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.553448 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="extract-utilities" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.553588 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f0d439-2c8c-42d1-8cdc-84df2101407f" containerName="registry-server" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.554622 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.556423 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l5bmb" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.566536 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh"] Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.640979 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.641025 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfd75\" (UniqueName: \"kubernetes.io/projected/e046d934-906d-4df1-adcb-f8d67d3d01df-kube-api-access-pfd75\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.641103 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.742439 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.742480 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfd75\" (UniqueName: \"kubernetes.io/projected/e046d934-906d-4df1-adcb-f8d67d3d01df-kube-api-access-pfd75\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.742507 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.742943 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.742977 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.764654 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfd75\" (UniqueName: \"kubernetes.io/projected/e046d934-906d-4df1-adcb-f8d67d3d01df-kube-api-access-pfd75\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:43 crc kubenswrapper[4914]: I1122 14:40:43.871151 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:44 crc kubenswrapper[4914]: I1122 14:40:44.320284 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh"] Nov 22 14:40:45 crc kubenswrapper[4914]: I1122 14:40:45.334638 4914 generic.go:334] "Generic (PLEG): container finished" podID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerID="05e0886c4bf12ced876e4252f90d67b7b6cc8756f2f22e89a033ece633b58f3c" exitCode=0 Nov 22 14:40:45 crc kubenswrapper[4914]: I1122 14:40:45.334742 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" event={"ID":"e046d934-906d-4df1-adcb-f8d67d3d01df","Type":"ContainerDied","Data":"05e0886c4bf12ced876e4252f90d67b7b6cc8756f2f22e89a033ece633b58f3c"} Nov 22 14:40:45 crc kubenswrapper[4914]: I1122 14:40:45.335068 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" event={"ID":"e046d934-906d-4df1-adcb-f8d67d3d01df","Type":"ContainerStarted","Data":"1b403584800b7bdd211e801d167011f049dec16343db19e4f4475f5d7f689ff7"} Nov 22 14:40:46 crc kubenswrapper[4914]: I1122 14:40:46.344563 4914 generic.go:334] "Generic (PLEG): container finished" podID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerID="48d495fceb8c31daa7c88193a99fd0be30a58bb3a67180d8a0bb946885e82ce3" exitCode=0 Nov 22 14:40:46 crc kubenswrapper[4914]: I1122 14:40:46.344753 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" event={"ID":"e046d934-906d-4df1-adcb-f8d67d3d01df","Type":"ContainerDied","Data":"48d495fceb8c31daa7c88193a99fd0be30a58bb3a67180d8a0bb946885e82ce3"} Nov 22 14:40:47 crc kubenswrapper[4914]: I1122 14:40:47.354061 4914 generic.go:334] "Generic (PLEG): container finished" podID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerID="baa3ba666fb98b5a467cfd9dd0a86db9bcf823608070d8bc15dc289160964a28" exitCode=0 Nov 22 14:40:47 crc kubenswrapper[4914]: I1122 14:40:47.354103 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" event={"ID":"e046d934-906d-4df1-adcb-f8d67d3d01df","Type":"ContainerDied","Data":"baa3ba666fb98b5a467cfd9dd0a86db9bcf823608070d8bc15dc289160964a28"} Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.799627 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.824945 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-util\") pod \"e046d934-906d-4df1-adcb-f8d67d3d01df\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.825020 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-bundle\") pod \"e046d934-906d-4df1-adcb-f8d67d3d01df\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.825227 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfd75\" (UniqueName: \"kubernetes.io/projected/e046d934-906d-4df1-adcb-f8d67d3d01df-kube-api-access-pfd75\") pod \"e046d934-906d-4df1-adcb-f8d67d3d01df\" (UID: \"e046d934-906d-4df1-adcb-f8d67d3d01df\") " Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.826932 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-bundle" (OuterVolumeSpecName: "bundle") pod "e046d934-906d-4df1-adcb-f8d67d3d01df" (UID: "e046d934-906d-4df1-adcb-f8d67d3d01df"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.831831 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e046d934-906d-4df1-adcb-f8d67d3d01df-kube-api-access-pfd75" (OuterVolumeSpecName: "kube-api-access-pfd75") pod "e046d934-906d-4df1-adcb-f8d67d3d01df" (UID: "e046d934-906d-4df1-adcb-f8d67d3d01df"). InnerVolumeSpecName "kube-api-access-pfd75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.846414 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-util" (OuterVolumeSpecName: "util") pod "e046d934-906d-4df1-adcb-f8d67d3d01df" (UID: "e046d934-906d-4df1-adcb-f8d67d3d01df"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.926917 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.926951 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e046d934-906d-4df1-adcb-f8d67d3d01df-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:48 crc kubenswrapper[4914]: I1122 14:40:48.926965 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfd75\" (UniqueName: \"kubernetes.io/projected/e046d934-906d-4df1-adcb-f8d67d3d01df-kube-api-access-pfd75\") on node \"crc\" DevicePath \"\"" Nov 22 14:40:49 crc kubenswrapper[4914]: I1122 14:40:49.374018 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" event={"ID":"e046d934-906d-4df1-adcb-f8d67d3d01df","Type":"ContainerDied","Data":"1b403584800b7bdd211e801d167011f049dec16343db19e4f4475f5d7f689ff7"} Nov 22 14:40:49 crc kubenswrapper[4914]: I1122 14:40:49.374081 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b403584800b7bdd211e801d167011f049dec16343db19e4f4475f5d7f689ff7" Nov 22 14:40:49 crc kubenswrapper[4914]: I1122 14:40:49.374249 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.916263 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p5cq7"] Nov 22 14:40:52 crc kubenswrapper[4914]: E1122 14:40:52.917215 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="util" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.917303 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="util" Nov 22 14:40:52 crc kubenswrapper[4914]: E1122 14:40:52.917368 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="extract" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.917416 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="extract" Nov 22 14:40:52 crc kubenswrapper[4914]: E1122 14:40:52.917467 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="pull" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.917535 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="pull" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.917700 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="e046d934-906d-4df1-adcb-f8d67d3d01df" containerName="extract" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.918540 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:52 crc kubenswrapper[4914]: I1122 14:40:52.936556 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5cq7"] Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.087298 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-catalog-content\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.087345 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-utilities\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.087452 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt8f6\" (UniqueName: \"kubernetes.io/projected/36b89328-04da-4861-b37c-05e945bad472-kube-api-access-kt8f6\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.188649 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt8f6\" (UniqueName: \"kubernetes.io/projected/36b89328-04da-4861-b37c-05e945bad472-kube-api-access-kt8f6\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.189053 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-catalog-content\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.189083 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-utilities\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.189630 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-utilities\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.189673 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-catalog-content\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.207960 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt8f6\" (UniqueName: \"kubernetes.io/projected/36b89328-04da-4861-b37c-05e945bad472-kube-api-access-kt8f6\") pod \"redhat-operators-p5cq7\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.233710 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.660848 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5cq7"] Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.912422 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qclh5"] Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.913469 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.927497 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qclh5"] Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.999010 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-utilities\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.999061 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72m2x\" (UniqueName: \"kubernetes.io/projected/5c5bdeb8-799d-482e-bfa8-73511bb00e97-kube-api-access-72m2x\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:53 crc kubenswrapper[4914]: I1122 14:40:53.999120 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-catalog-content\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.100548 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-catalog-content\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.100635 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-utilities\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.100668 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72m2x\" (UniqueName: \"kubernetes.io/projected/5c5bdeb8-799d-482e-bfa8-73511bb00e97-kube-api-access-72m2x\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.101253 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-catalog-content\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.101323 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-utilities\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.117161 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72m2x\" (UniqueName: \"kubernetes.io/projected/5c5bdeb8-799d-482e-bfa8-73511bb00e97-kube-api-access-72m2x\") pod \"redhat-marketplace-qclh5\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.226316 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.410826 4914 generic.go:334] "Generic (PLEG): container finished" podID="36b89328-04da-4861-b37c-05e945bad472" containerID="e52426445593c76f45c978236249effd59184a649a3fbf8d0ffde2fe568b357e" exitCode=0 Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.411018 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerDied","Data":"e52426445593c76f45c978236249effd59184a649a3fbf8d0ffde2fe568b357e"} Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.411097 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerStarted","Data":"3deaa93325423b0fd312296e2722b1b39c77dcf77e2f056f0c6aa1f2c7a3d771"} Nov 22 14:40:54 crc kubenswrapper[4914]: I1122 14:40:54.602647 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qclh5"] Nov 22 14:40:54 crc kubenswrapper[4914]: W1122 14:40:54.608030 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c5bdeb8_799d_482e_bfa8_73511bb00e97.slice/crio-85365a5f55bc2672603bd2fb7c6dc59aa26b90805248e0bdea7f5342647139d1 WatchSource:0}: Error finding container 85365a5f55bc2672603bd2fb7c6dc59aa26b90805248e0bdea7f5342647139d1: Status 404 returned error can't find the container with id 85365a5f55bc2672603bd2fb7c6dc59aa26b90805248e0bdea7f5342647139d1 Nov 22 14:40:55 crc kubenswrapper[4914]: I1122 14:40:55.420862 4914 generic.go:334] "Generic (PLEG): container finished" podID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerID="4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779" exitCode=0 Nov 22 14:40:55 crc kubenswrapper[4914]: I1122 14:40:55.421081 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qclh5" event={"ID":"5c5bdeb8-799d-482e-bfa8-73511bb00e97","Type":"ContainerDied","Data":"4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779"} Nov 22 14:40:55 crc kubenswrapper[4914]: I1122 14:40:55.421409 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qclh5" event={"ID":"5c5bdeb8-799d-482e-bfa8-73511bb00e97","Type":"ContainerStarted","Data":"85365a5f55bc2672603bd2fb7c6dc59aa26b90805248e0bdea7f5342647139d1"} Nov 22 14:40:55 crc kubenswrapper[4914]: I1122 14:40:55.428019 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerStarted","Data":"ddb74e899dd5da4394f9d1c5b5c73ca52bba1ffb7bb4e8ed7f391e8a9a685c33"} Nov 22 14:40:56 crc kubenswrapper[4914]: I1122 14:40:56.436447 4914 generic.go:334] "Generic (PLEG): container finished" podID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerID="0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637" exitCode=0 Nov 22 14:40:56 crc kubenswrapper[4914]: I1122 14:40:56.436562 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qclh5" event={"ID":"5c5bdeb8-799d-482e-bfa8-73511bb00e97","Type":"ContainerDied","Data":"0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637"} Nov 22 14:40:56 crc kubenswrapper[4914]: I1122 14:40:56.438630 4914 generic.go:334] "Generic (PLEG): container finished" podID="36b89328-04da-4861-b37c-05e945bad472" containerID="ddb74e899dd5da4394f9d1c5b5c73ca52bba1ffb7bb4e8ed7f391e8a9a685c33" exitCode=0 Nov 22 14:40:56 crc kubenswrapper[4914]: I1122 14:40:56.438670 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerDied","Data":"ddb74e899dd5da4394f9d1c5b5c73ca52bba1ffb7bb4e8ed7f391e8a9a685c33"} Nov 22 14:40:57 crc kubenswrapper[4914]: I1122 14:40:57.445541 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qclh5" event={"ID":"5c5bdeb8-799d-482e-bfa8-73511bb00e97","Type":"ContainerStarted","Data":"ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea"} Nov 22 14:40:57 crc kubenswrapper[4914]: I1122 14:40:57.448347 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerStarted","Data":"cb1154e9ba5368a5ec609fc0f99acd3d993bd6c5f200ee1e9fa7ee22767c8a6f"} Nov 22 14:40:57 crc kubenswrapper[4914]: I1122 14:40:57.466131 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qclh5" podStartSLOduration=3.053204267 podStartE2EDuration="4.466116034s" podCreationTimestamp="2025-11-22 14:40:53 +0000 UTC" firstStartedPulling="2025-11-22 14:40:55.423235701 +0000 UTC m=+821.231805402" lastFinishedPulling="2025-11-22 14:40:56.836147498 +0000 UTC m=+822.644717169" observedRunningTime="2025-11-22 14:40:57.463982441 +0000 UTC m=+823.272552102" watchObservedRunningTime="2025-11-22 14:40:57.466116034 +0000 UTC m=+823.274685695" Nov 22 14:40:57 crc kubenswrapper[4914]: I1122 14:40:57.482472 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p5cq7" podStartSLOduration=3.087447746 podStartE2EDuration="5.482456892s" podCreationTimestamp="2025-11-22 14:40:52 +0000 UTC" firstStartedPulling="2025-11-22 14:40:54.41240076 +0000 UTC m=+820.220970421" lastFinishedPulling="2025-11-22 14:40:56.807409896 +0000 UTC m=+822.615979567" observedRunningTime="2025-11-22 14:40:57.479401248 +0000 UTC m=+823.287970899" watchObservedRunningTime="2025-11-22 14:40:57.482456892 +0000 UTC m=+823.291026553" Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.545029 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z"] Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.546774 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.552289 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-rhztr" Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.559049 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z"] Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.670151 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh55g\" (UniqueName: \"kubernetes.io/projected/8b4d9e00-9685-4ca8-8b51-4246e6e65d65-kube-api-access-bh55g\") pod \"rabbitmq-cluster-operator-779fc9694b-qdh2z\" (UID: \"8b4d9e00-9685-4ca8-8b51-4246e6e65d65\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.772286 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh55g\" (UniqueName: \"kubernetes.io/projected/8b4d9e00-9685-4ca8-8b51-4246e6e65d65-kube-api-access-bh55g\") pod \"rabbitmq-cluster-operator-779fc9694b-qdh2z\" (UID: \"8b4d9e00-9685-4ca8-8b51-4246e6e65d65\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.799517 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh55g\" (UniqueName: \"kubernetes.io/projected/8b4d9e00-9685-4ca8-8b51-4246e6e65d65-kube-api-access-bh55g\") pod \"rabbitmq-cluster-operator-779fc9694b-qdh2z\" (UID: \"8b4d9e00-9685-4ca8-8b51-4246e6e65d65\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" Nov 22 14:40:59 crc kubenswrapper[4914]: I1122 14:40:59.861478 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" Nov 22 14:41:00 crc kubenswrapper[4914]: I1122 14:41:00.328674 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z"] Nov 22 14:41:00 crc kubenswrapper[4914]: I1122 14:41:00.477285 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" event={"ID":"8b4d9e00-9685-4ca8-8b51-4246e6e65d65","Type":"ContainerStarted","Data":"c521eebb59df38f9d6d402faa7980193df58762533dead32d521e753d738662a"} Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.327546 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b2h25"] Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.328973 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.337590 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b2h25"] Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.419115 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq44k\" (UniqueName: \"kubernetes.io/projected/b57aafee-5fcd-494a-a60a-effd73c75644-kube-api-access-wq44k\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.419329 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-catalog-content\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.419381 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-utilities\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.519825 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq44k\" (UniqueName: \"kubernetes.io/projected/b57aafee-5fcd-494a-a60a-effd73c75644-kube-api-access-wq44k\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.519949 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-catalog-content\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.519977 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-utilities\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.520440 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-utilities\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.520531 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-catalog-content\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.540759 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq44k\" (UniqueName: \"kubernetes.io/projected/b57aafee-5fcd-494a-a60a-effd73c75644-kube-api-access-wq44k\") pod \"certified-operators-b2h25\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:02 crc kubenswrapper[4914]: I1122 14:41:02.655381 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:03 crc kubenswrapper[4914]: I1122 14:41:03.234645 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:41:03 crc kubenswrapper[4914]: I1122 14:41:03.234703 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:41:03 crc kubenswrapper[4914]: I1122 14:41:03.271526 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:41:03 crc kubenswrapper[4914]: I1122 14:41:03.547425 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.221128 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b2h25"] Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.230707 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.230752 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:41:04 crc kubenswrapper[4914]: W1122 14:41:04.230826 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb57aafee_5fcd_494a_a60a_effd73c75644.slice/crio-8b2edab3795b4f21915b0afc309313fd792f619defd90f6f58e6bd23dfd20739 WatchSource:0}: Error finding container 8b2edab3795b4f21915b0afc309313fd792f619defd90f6f58e6bd23dfd20739: Status 404 returned error can't find the container with id 8b2edab3795b4f21915b0afc309313fd792f619defd90f6f58e6bd23dfd20739 Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.288095 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.499038 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2h25" event={"ID":"b57aafee-5fcd-494a-a60a-effd73c75644","Type":"ContainerStarted","Data":"8b2edab3795b4f21915b0afc309313fd792f619defd90f6f58e6bd23dfd20739"} Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.500355 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" event={"ID":"8b4d9e00-9685-4ca8-8b51-4246e6e65d65","Type":"ContainerStarted","Data":"d693253ec2d37d5365576d2173fc6c7c287fac3ce5056d6d965e3532ca9d0000"} Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.519289 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qdh2z" podStartSLOduration=1.956836783 podStartE2EDuration="5.519268453s" podCreationTimestamp="2025-11-22 14:40:59 +0000 UTC" firstStartedPulling="2025-11-22 14:41:00.342571241 +0000 UTC m=+826.151140902" lastFinishedPulling="2025-11-22 14:41:03.905002911 +0000 UTC m=+829.713572572" observedRunningTime="2025-11-22 14:41:04.516295391 +0000 UTC m=+830.324865072" watchObservedRunningTime="2025-11-22 14:41:04.519268453 +0000 UTC m=+830.327838114" Nov 22 14:41:04 crc kubenswrapper[4914]: I1122 14:41:04.544159 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:41:05 crc kubenswrapper[4914]: I1122 14:41:05.506633 4914 generic.go:334] "Generic (PLEG): container finished" podID="b57aafee-5fcd-494a-a60a-effd73c75644" containerID="0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc" exitCode=0 Nov 22 14:41:05 crc kubenswrapper[4914]: I1122 14:41:05.506674 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2h25" event={"ID":"b57aafee-5fcd-494a-a60a-effd73c75644","Type":"ContainerDied","Data":"0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc"} Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.104509 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5cq7"] Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.105023 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p5cq7" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="registry-server" containerID="cri-o://cb1154e9ba5368a5ec609fc0f99acd3d993bd6c5f200ee1e9fa7ee22767c8a6f" gracePeriod=2 Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.520430 4914 generic.go:334] "Generic (PLEG): container finished" podID="36b89328-04da-4861-b37c-05e945bad472" containerID="cb1154e9ba5368a5ec609fc0f99acd3d993bd6c5f200ee1e9fa7ee22767c8a6f" exitCode=0 Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.520500 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerDied","Data":"cb1154e9ba5368a5ec609fc0f99acd3d993bd6c5f200ee1e9fa7ee22767c8a6f"} Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.520542 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5cq7" event={"ID":"36b89328-04da-4861-b37c-05e945bad472","Type":"ContainerDied","Data":"3deaa93325423b0fd312296e2722b1b39c77dcf77e2f056f0c6aa1f2c7a3d771"} Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.520555 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3deaa93325423b0fd312296e2722b1b39c77dcf77e2f056f0c6aa1f2c7a3d771" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.521937 4914 generic.go:334] "Generic (PLEG): container finished" podID="b57aafee-5fcd-494a-a60a-effd73c75644" containerID="95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0" exitCode=0 Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.521987 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2h25" event={"ID":"b57aafee-5fcd-494a-a60a-effd73c75644","Type":"ContainerDied","Data":"95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0"} Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.530617 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.689107 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt8f6\" (UniqueName: \"kubernetes.io/projected/36b89328-04da-4861-b37c-05e945bad472-kube-api-access-kt8f6\") pod \"36b89328-04da-4861-b37c-05e945bad472\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.689293 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-utilities\") pod \"36b89328-04da-4861-b37c-05e945bad472\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.689333 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-catalog-content\") pod \"36b89328-04da-4861-b37c-05e945bad472\" (UID: \"36b89328-04da-4861-b37c-05e945bad472\") " Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.690098 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-utilities" (OuterVolumeSpecName: "utilities") pod "36b89328-04da-4861-b37c-05e945bad472" (UID: "36b89328-04da-4861-b37c-05e945bad472"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.697170 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b89328-04da-4861-b37c-05e945bad472-kube-api-access-kt8f6" (OuterVolumeSpecName: "kube-api-access-kt8f6") pod "36b89328-04da-4861-b37c-05e945bad472" (UID: "36b89328-04da-4861-b37c-05e945bad472"). InnerVolumeSpecName "kube-api-access-kt8f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.782757 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36b89328-04da-4861-b37c-05e945bad472" (UID: "36b89328-04da-4861-b37c-05e945bad472"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.790588 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.790657 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b89328-04da-4861-b37c-05e945bad472-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:07 crc kubenswrapper[4914]: I1122 14:41:07.790668 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt8f6\" (UniqueName: \"kubernetes.io/projected/36b89328-04da-4861-b37c-05e945bad472-kube-api-access-kt8f6\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:08 crc kubenswrapper[4914]: I1122 14:41:08.533066 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5cq7" Nov 22 14:41:08 crc kubenswrapper[4914]: I1122 14:41:08.533156 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2h25" event={"ID":"b57aafee-5fcd-494a-a60a-effd73c75644","Type":"ContainerStarted","Data":"70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb"} Nov 22 14:41:08 crc kubenswrapper[4914]: I1122 14:41:08.571194 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b2h25" podStartSLOduration=3.975105816 podStartE2EDuration="6.5711719s" podCreationTimestamp="2025-11-22 14:41:02 +0000 UTC" firstStartedPulling="2025-11-22 14:41:05.508311513 +0000 UTC m=+831.316881174" lastFinishedPulling="2025-11-22 14:41:08.104377597 +0000 UTC m=+833.912947258" observedRunningTime="2025-11-22 14:41:08.556347098 +0000 UTC m=+834.364916789" watchObservedRunningTime="2025-11-22 14:41:08.5711719 +0000 UTC m=+834.379741561" Nov 22 14:41:08 crc kubenswrapper[4914]: I1122 14:41:08.575328 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5cq7"] Nov 22 14:41:08 crc kubenswrapper[4914]: I1122 14:41:08.581451 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p5cq7"] Nov 22 14:41:08 crc kubenswrapper[4914]: I1122 14:41:08.875496 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b89328-04da-4861-b37c-05e945bad472" path="/var/lib/kubelet/pods/36b89328-04da-4861-b37c-05e945bad472/volumes" Nov 22 14:41:09 crc kubenswrapper[4914]: I1122 14:41:09.701164 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qclh5"] Nov 22 14:41:09 crc kubenswrapper[4914]: I1122 14:41:09.701389 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qclh5" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="registry-server" containerID="cri-o://ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea" gracePeriod=2 Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.118551 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 22 14:41:10 crc kubenswrapper[4914]: E1122 14:41:10.119319 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="extract-utilities" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.119350 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="extract-utilities" Nov 22 14:41:10 crc kubenswrapper[4914]: E1122 14:41:10.119386 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="extract-content" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.119401 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="extract-content" Nov 22 14:41:10 crc kubenswrapper[4914]: E1122 14:41:10.119424 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="registry-server" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.119438 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="registry-server" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.119647 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b89328-04da-4861-b37c-05e945bad472" containerName="registry-server" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.120869 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.123459 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-qk9kx" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.123503 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.123613 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.126034 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.127548 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.141765 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.161583 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226521 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226586 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226617 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226653 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226743 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226766 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226799 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-196e3177-0b52-4674-a327-ded6b3769409\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196e3177-0b52-4674-a327-ded6b3769409\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.226822 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqng\" (UniqueName: \"kubernetes.io/projected/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-kube-api-access-zsqng\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.327994 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-catalog-content\") pod \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328060 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72m2x\" (UniqueName: \"kubernetes.io/projected/5c5bdeb8-799d-482e-bfa8-73511bb00e97-kube-api-access-72m2x\") pod \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328081 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-utilities\") pod \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\" (UID: \"5c5bdeb8-799d-482e-bfa8-73511bb00e97\") " Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328319 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328345 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328372 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-196e3177-0b52-4674-a327-ded6b3769409\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196e3177-0b52-4674-a327-ded6b3769409\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328391 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqng\" (UniqueName: \"kubernetes.io/projected/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-kube-api-access-zsqng\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328421 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328438 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328458 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.328481 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.335974 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c5bdeb8-799d-482e-bfa8-73511bb00e97-kube-api-access-72m2x" (OuterVolumeSpecName: "kube-api-access-72m2x") pod "5c5bdeb8-799d-482e-bfa8-73511bb00e97" (UID: "5c5bdeb8-799d-482e-bfa8-73511bb00e97"). InnerVolumeSpecName "kube-api-access-72m2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.336519 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-utilities" (OuterVolumeSpecName: "utilities") pod "5c5bdeb8-799d-482e-bfa8-73511bb00e97" (UID: "5c5bdeb8-799d-482e-bfa8-73511bb00e97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.336850 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.337262 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.337688 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.338188 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.338991 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.339493 4914 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.339518 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-196e3177-0b52-4674-a327-ded6b3769409\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196e3177-0b52-4674-a327-ded6b3769409\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ecc69bd981b6befb70eca5a815c86f6ab4bf6f699442b18bbf0b1fdfc17956cc/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.344221 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.360360 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c5bdeb8-799d-482e-bfa8-73511bb00e97" (UID: "5c5bdeb8-799d-482e-bfa8-73511bb00e97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.367015 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqng\" (UniqueName: \"kubernetes.io/projected/713f96a8-d1f3-4fba-bd02-f025ac07a4b1-kube-api-access-zsqng\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.376096 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-196e3177-0b52-4674-a327-ded6b3769409\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-196e3177-0b52-4674-a327-ded6b3769409\") pod \"rabbitmq-server-0\" (UID: \"713f96a8-d1f3-4fba-bd02-f025ac07a4b1\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.430527 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.430567 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72m2x\" (UniqueName: \"kubernetes.io/projected/5c5bdeb8-799d-482e-bfa8-73511bb00e97-kube-api-access-72m2x\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.430579 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c5bdeb8-799d-482e-bfa8-73511bb00e97-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.474204 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.566913 4914 generic.go:334] "Generic (PLEG): container finished" podID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerID="ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea" exitCode=0 Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.567208 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qclh5" event={"ID":"5c5bdeb8-799d-482e-bfa8-73511bb00e97","Type":"ContainerDied","Data":"ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea"} Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.567233 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qclh5" event={"ID":"5c5bdeb8-799d-482e-bfa8-73511bb00e97","Type":"ContainerDied","Data":"85365a5f55bc2672603bd2fb7c6dc59aa26b90805248e0bdea7f5342647139d1"} Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.567249 4914 scope.go:117] "RemoveContainer" containerID="ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.567369 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qclh5" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.588693 4914 scope.go:117] "RemoveContainer" containerID="0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.601754 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qclh5"] Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.603173 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qclh5"] Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.621463 4914 scope.go:117] "RemoveContainer" containerID="4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.640215 4914 scope.go:117] "RemoveContainer" containerID="ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea" Nov 22 14:41:10 crc kubenswrapper[4914]: E1122 14:41:10.640910 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea\": container with ID starting with ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea not found: ID does not exist" containerID="ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.640946 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea"} err="failed to get container status \"ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea\": rpc error: code = NotFound desc = could not find container \"ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea\": container with ID starting with ac8bb5914c6ea4d57a0e46bdf33e0d6e5e4ccaf57611f1771d7b6aa7312fe7ea not found: ID does not exist" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.640970 4914 scope.go:117] "RemoveContainer" containerID="0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637" Nov 22 14:41:10 crc kubenswrapper[4914]: E1122 14:41:10.641255 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637\": container with ID starting with 0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637 not found: ID does not exist" containerID="0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.641307 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637"} err="failed to get container status \"0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637\": rpc error: code = NotFound desc = could not find container \"0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637\": container with ID starting with 0519fa3280c0bb85a8a9fe7781a70d7ab6dc23121fc1b39db3a611485fd10637 not found: ID does not exist" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.641327 4914 scope.go:117] "RemoveContainer" containerID="4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779" Nov 22 14:41:10 crc kubenswrapper[4914]: E1122 14:41:10.641573 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779\": container with ID starting with 4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779 not found: ID does not exist" containerID="4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.641612 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779"} err="failed to get container status \"4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779\": rpc error: code = NotFound desc = could not find container \"4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779\": container with ID starting with 4458412916bb26a89ddcedc1ebf3ffd9e38fcc9e3de363afb1878893c5545779 not found: ID does not exist" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.877520 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" path="/var/lib/kubelet/pods/5c5bdeb8-799d-482e-bfa8-73511bb00e97/volumes" Nov 22 14:41:10 crc kubenswrapper[4914]: I1122 14:41:10.889505 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 22 14:41:11 crc kubenswrapper[4914]: I1122 14:41:11.575780 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"713f96a8-d1f3-4fba-bd02-f025ac07a4b1","Type":"ContainerStarted","Data":"29043d19d3c294de2df5cd2dcf42fd588dccc093f7b9b7a0be4d9c916e83cc04"} Nov 22 14:41:12 crc kubenswrapper[4914]: I1122 14:41:12.656918 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:12 crc kubenswrapper[4914]: I1122 14:41:12.656976 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:12 crc kubenswrapper[4914]: I1122 14:41:12.704579 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:13 crc kubenswrapper[4914]: I1122 14:41:13.669600 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.109029 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-zs5qg"] Nov 22 14:41:15 crc kubenswrapper[4914]: E1122 14:41:15.110811 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="registry-server" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.110897 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="registry-server" Nov 22 14:41:15 crc kubenswrapper[4914]: E1122 14:41:15.110957 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="extract-content" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.111007 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="extract-content" Nov 22 14:41:15 crc kubenswrapper[4914]: E1122 14:41:15.111065 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="extract-utilities" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.111141 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="extract-utilities" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.111308 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c5bdeb8-799d-482e-bfa8-73511bb00e97" containerName="registry-server" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.111740 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.113972 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-sr72l" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.120246 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-zs5qg"] Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.238250 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxpf5\" (UniqueName: \"kubernetes.io/projected/8ef98c25-d009-447a-a560-94871072156d-kube-api-access-vxpf5\") pod \"keystone-operator-index-zs5qg\" (UID: \"8ef98c25-d009-447a-a560-94871072156d\") " pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.339699 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxpf5\" (UniqueName: \"kubernetes.io/projected/8ef98c25-d009-447a-a560-94871072156d-kube-api-access-vxpf5\") pod \"keystone-operator-index-zs5qg\" (UID: \"8ef98c25-d009-447a-a560-94871072156d\") " pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.357602 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxpf5\" (UniqueName: \"kubernetes.io/projected/8ef98c25-d009-447a-a560-94871072156d-kube-api-access-vxpf5\") pod \"keystone-operator-index-zs5qg\" (UID: \"8ef98c25-d009-447a-a560-94871072156d\") " pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:15 crc kubenswrapper[4914]: I1122 14:41:15.426538 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:16 crc kubenswrapper[4914]: I1122 14:41:16.406464 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-zs5qg"] Nov 22 14:41:16 crc kubenswrapper[4914]: W1122 14:41:16.419491 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ef98c25_d009_447a_a560_94871072156d.slice/crio-e139ac1514f9dbfeeb99854682c44b90980ed90f75e616e0beaa9748a916bd72 WatchSource:0}: Error finding container e139ac1514f9dbfeeb99854682c44b90980ed90f75e616e0beaa9748a916bd72: Status 404 returned error can't find the container with id e139ac1514f9dbfeeb99854682c44b90980ed90f75e616e0beaa9748a916bd72 Nov 22 14:41:16 crc kubenswrapper[4914]: I1122 14:41:16.648015 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-zs5qg" event={"ID":"8ef98c25-d009-447a-a560-94871072156d","Type":"ContainerStarted","Data":"e139ac1514f9dbfeeb99854682c44b90980ed90f75e616e0beaa9748a916bd72"} Nov 22 14:41:17 crc kubenswrapper[4914]: I1122 14:41:17.658207 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"713f96a8-d1f3-4fba-bd02-f025ac07a4b1","Type":"ContainerStarted","Data":"a81a6a1ee0a34f46d7875bcab79eac880ad13e2adaeb76edf7a7144f937921f7"} Nov 22 14:41:17 crc kubenswrapper[4914]: I1122 14:41:17.664817 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-zs5qg" event={"ID":"8ef98c25-d009-447a-a560-94871072156d","Type":"ContainerStarted","Data":"942224ded5949cffd05a37d833d0ab2c07e49a9491abd7c9e6ca2271abb3f176"} Nov 22 14:41:17 crc kubenswrapper[4914]: I1122 14:41:17.711435 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-zs5qg" podStartSLOduration=1.875348424 podStartE2EDuration="2.71140867s" podCreationTimestamp="2025-11-22 14:41:15 +0000 UTC" firstStartedPulling="2025-11-22 14:41:16.423197448 +0000 UTC m=+842.231767109" lastFinishedPulling="2025-11-22 14:41:17.259257684 +0000 UTC m=+843.067827355" observedRunningTime="2025-11-22 14:41:17.703276252 +0000 UTC m=+843.511845953" watchObservedRunningTime="2025-11-22 14:41:17.71140867 +0000 UTC m=+843.519978371" Nov 22 14:41:20 crc kubenswrapper[4914]: I1122 14:41:20.702514 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b2h25"] Nov 22 14:41:20 crc kubenswrapper[4914]: I1122 14:41:20.703182 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b2h25" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="registry-server" containerID="cri-o://70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb" gracePeriod=2 Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.248160 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.437725 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq44k\" (UniqueName: \"kubernetes.io/projected/b57aafee-5fcd-494a-a60a-effd73c75644-kube-api-access-wq44k\") pod \"b57aafee-5fcd-494a-a60a-effd73c75644\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.438003 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-catalog-content\") pod \"b57aafee-5fcd-494a-a60a-effd73c75644\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.438268 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-utilities\") pod \"b57aafee-5fcd-494a-a60a-effd73c75644\" (UID: \"b57aafee-5fcd-494a-a60a-effd73c75644\") " Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.439399 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-utilities" (OuterVolumeSpecName: "utilities") pod "b57aafee-5fcd-494a-a60a-effd73c75644" (UID: "b57aafee-5fcd-494a-a60a-effd73c75644"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.446101 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b57aafee-5fcd-494a-a60a-effd73c75644-kube-api-access-wq44k" (OuterVolumeSpecName: "kube-api-access-wq44k") pod "b57aafee-5fcd-494a-a60a-effd73c75644" (UID: "b57aafee-5fcd-494a-a60a-effd73c75644"). InnerVolumeSpecName "kube-api-access-wq44k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.516471 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b57aafee-5fcd-494a-a60a-effd73c75644" (UID: "b57aafee-5fcd-494a-a60a-effd73c75644"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.540193 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.540220 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b57aafee-5fcd-494a-a60a-effd73c75644-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.540234 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq44k\" (UniqueName: \"kubernetes.io/projected/b57aafee-5fcd-494a-a60a-effd73c75644-kube-api-access-wq44k\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.700085 4914 generic.go:334] "Generic (PLEG): container finished" podID="b57aafee-5fcd-494a-a60a-effd73c75644" containerID="70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb" exitCode=0 Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.700136 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2h25" event={"ID":"b57aafee-5fcd-494a-a60a-effd73c75644","Type":"ContainerDied","Data":"70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb"} Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.700165 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2h25" event={"ID":"b57aafee-5fcd-494a-a60a-effd73c75644","Type":"ContainerDied","Data":"8b2edab3795b4f21915b0afc309313fd792f619defd90f6f58e6bd23dfd20739"} Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.700171 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b2h25" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.700186 4914 scope.go:117] "RemoveContainer" containerID="70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.732193 4914 scope.go:117] "RemoveContainer" containerID="95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.746862 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b2h25"] Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.753151 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b2h25"] Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.777235 4914 scope.go:117] "RemoveContainer" containerID="0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.823361 4914 scope.go:117] "RemoveContainer" containerID="70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb" Nov 22 14:41:21 crc kubenswrapper[4914]: E1122 14:41:21.824301 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb\": container with ID starting with 70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb not found: ID does not exist" containerID="70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.824380 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb"} err="failed to get container status \"70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb\": rpc error: code = NotFound desc = could not find container \"70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb\": container with ID starting with 70599d3e19169ef59662781d0cfc068a9dfc920dda6debb0ff18365cfc3bbbfb not found: ID does not exist" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.824438 4914 scope.go:117] "RemoveContainer" containerID="95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0" Nov 22 14:41:21 crc kubenswrapper[4914]: E1122 14:41:21.825174 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0\": container with ID starting with 95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0 not found: ID does not exist" containerID="95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.825232 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0"} err="failed to get container status \"95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0\": rpc error: code = NotFound desc = could not find container \"95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0\": container with ID starting with 95d0fad562818767b1026134e58d668d5270103318150d20e7932457031986a0 not found: ID does not exist" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.825278 4914 scope.go:117] "RemoveContainer" containerID="0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc" Nov 22 14:41:21 crc kubenswrapper[4914]: E1122 14:41:21.825869 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc\": container with ID starting with 0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc not found: ID does not exist" containerID="0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc" Nov 22 14:41:21 crc kubenswrapper[4914]: I1122 14:41:21.825992 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc"} err="failed to get container status \"0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc\": rpc error: code = NotFound desc = could not find container \"0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc\": container with ID starting with 0295d40711d75fbf922b9a5fcee7074972ebbb8297bfeeef6b6612709681d4fc not found: ID does not exist" Nov 22 14:41:22 crc kubenswrapper[4914]: I1122 14:41:22.877422 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" path="/var/lib/kubelet/pods/b57aafee-5fcd-494a-a60a-effd73c75644/volumes" Nov 22 14:41:25 crc kubenswrapper[4914]: I1122 14:41:25.426804 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:25 crc kubenswrapper[4914]: I1122 14:41:25.427092 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:25 crc kubenswrapper[4914]: I1122 14:41:25.463094 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:25 crc kubenswrapper[4914]: I1122 14:41:25.764494 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-zs5qg" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.167162 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms"] Nov 22 14:41:29 crc kubenswrapper[4914]: E1122 14:41:29.167830 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="extract-content" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.167850 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="extract-content" Nov 22 14:41:29 crc kubenswrapper[4914]: E1122 14:41:29.167865 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="extract-utilities" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.167878 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="extract-utilities" Nov 22 14:41:29 crc kubenswrapper[4914]: E1122 14:41:29.167909 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="registry-server" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.167917 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="registry-server" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.168052 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="b57aafee-5fcd-494a-a60a-effd73c75644" containerName="registry-server" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.169141 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.172275 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l5bmb" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.192405 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms"] Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.244141 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-util\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.244232 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkzrn\" (UniqueName: \"kubernetes.io/projected/992a8cd6-c0fc-42fd-9668-32260b1e9984-kube-api-access-tkzrn\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.244315 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-bundle\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.345461 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-util\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.345755 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkzrn\" (UniqueName: \"kubernetes.io/projected/992a8cd6-c0fc-42fd-9668-32260b1e9984-kube-api-access-tkzrn\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.346011 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-bundle\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.346448 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-util\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.346595 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-bundle\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.391174 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkzrn\" (UniqueName: \"kubernetes.io/projected/992a8cd6-c0fc-42fd-9668-32260b1e9984-kube-api-access-tkzrn\") pod \"e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.498741 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:29 crc kubenswrapper[4914]: I1122 14:41:29.914366 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms"] Nov 22 14:41:30 crc kubenswrapper[4914]: I1122 14:41:30.767381 4914 generic.go:334] "Generic (PLEG): container finished" podID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerID="4b23186f4faf53baa096da2efee6487167732524e7250442972804d7bc7438c5" exitCode=0 Nov 22 14:41:30 crc kubenswrapper[4914]: I1122 14:41:30.767439 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" event={"ID":"992a8cd6-c0fc-42fd-9668-32260b1e9984","Type":"ContainerDied","Data":"4b23186f4faf53baa096da2efee6487167732524e7250442972804d7bc7438c5"} Nov 22 14:41:30 crc kubenswrapper[4914]: I1122 14:41:30.767664 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" event={"ID":"992a8cd6-c0fc-42fd-9668-32260b1e9984","Type":"ContainerStarted","Data":"934c3cf4081abe6accf1594603302efed23fab6d24764a01f1855bd2e74c536b"} Nov 22 14:41:31 crc kubenswrapper[4914]: I1122 14:41:31.777607 4914 generic.go:334] "Generic (PLEG): container finished" podID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerID="afd3d55fa7dd35bd5e3f8517a49b00968ccbf8a0bf5a598b45fafa5d85b53c8a" exitCode=0 Nov 22 14:41:31 crc kubenswrapper[4914]: I1122 14:41:31.777870 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" event={"ID":"992a8cd6-c0fc-42fd-9668-32260b1e9984","Type":"ContainerDied","Data":"afd3d55fa7dd35bd5e3f8517a49b00968ccbf8a0bf5a598b45fafa5d85b53c8a"} Nov 22 14:41:32 crc kubenswrapper[4914]: I1122 14:41:32.789725 4914 generic.go:334] "Generic (PLEG): container finished" podID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerID="b75d30cec31191b2bc0364efb34f9bcb8c8647339ce381abf88a83cdd4e7b342" exitCode=0 Nov 22 14:41:32 crc kubenswrapper[4914]: I1122 14:41:32.789813 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" event={"ID":"992a8cd6-c0fc-42fd-9668-32260b1e9984","Type":"ContainerDied","Data":"b75d30cec31191b2bc0364efb34f9bcb8c8647339ce381abf88a83cdd4e7b342"} Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.169496 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.248344 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-bundle\") pod \"992a8cd6-c0fc-42fd-9668-32260b1e9984\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.248501 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-util\") pod \"992a8cd6-c0fc-42fd-9668-32260b1e9984\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.248547 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkzrn\" (UniqueName: \"kubernetes.io/projected/992a8cd6-c0fc-42fd-9668-32260b1e9984-kube-api-access-tkzrn\") pod \"992a8cd6-c0fc-42fd-9668-32260b1e9984\" (UID: \"992a8cd6-c0fc-42fd-9668-32260b1e9984\") " Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.249108 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-bundle" (OuterVolumeSpecName: "bundle") pod "992a8cd6-c0fc-42fd-9668-32260b1e9984" (UID: "992a8cd6-c0fc-42fd-9668-32260b1e9984"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.249362 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.257119 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992a8cd6-c0fc-42fd-9668-32260b1e9984-kube-api-access-tkzrn" (OuterVolumeSpecName: "kube-api-access-tkzrn") pod "992a8cd6-c0fc-42fd-9668-32260b1e9984" (UID: "992a8cd6-c0fc-42fd-9668-32260b1e9984"). InnerVolumeSpecName "kube-api-access-tkzrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.271555 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-util" (OuterVolumeSpecName: "util") pod "992a8cd6-c0fc-42fd-9668-32260b1e9984" (UID: "992a8cd6-c0fc-42fd-9668-32260b1e9984"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.350474 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/992a8cd6-c0fc-42fd-9668-32260b1e9984-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.350511 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkzrn\" (UniqueName: \"kubernetes.io/projected/992a8cd6-c0fc-42fd-9668-32260b1e9984-kube-api-access-tkzrn\") on node \"crc\" DevicePath \"\"" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.809437 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" event={"ID":"992a8cd6-c0fc-42fd-9668-32260b1e9984","Type":"ContainerDied","Data":"934c3cf4081abe6accf1594603302efed23fab6d24764a01f1855bd2e74c536b"} Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.809506 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms" Nov 22 14:41:34 crc kubenswrapper[4914]: I1122 14:41:34.809511 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="934c3cf4081abe6accf1594603302efed23fab6d24764a01f1855bd2e74c536b" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.761405 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj"] Nov 22 14:41:45 crc kubenswrapper[4914]: E1122 14:41:45.762254 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="extract" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.762269 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="extract" Nov 22 14:41:45 crc kubenswrapper[4914]: E1122 14:41:45.762282 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="util" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.762289 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="util" Nov 22 14:41:45 crc kubenswrapper[4914]: E1122 14:41:45.762308 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="pull" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.762317 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="pull" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.762466 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="992a8cd6-c0fc-42fd-9668-32260b1e9984" containerName="extract" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.763287 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.765490 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.777386 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ssxfj" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.780805 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj"] Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.921469 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-webhook-cert\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.921809 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-apiservice-cert\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:45 crc kubenswrapper[4914]: I1122 14:41:45.921862 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5rs\" (UniqueName: \"kubernetes.io/projected/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-kube-api-access-xn5rs\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.023366 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5rs\" (UniqueName: \"kubernetes.io/projected/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-kube-api-access-xn5rs\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.023444 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-webhook-cert\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.023507 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-apiservice-cert\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.029523 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-apiservice-cert\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.032444 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-webhook-cert\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.044659 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5rs\" (UniqueName: \"kubernetes.io/projected/94d945b4-bca0-45e6-8c5f-c2e471e2aad5-kube-api-access-xn5rs\") pod \"keystone-operator-controller-manager-676b6b6b4-h8ksj\" (UID: \"94d945b4-bca0-45e6-8c5f-c2e471e2aad5\") " pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.080760 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.497888 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj"] Nov 22 14:41:46 crc kubenswrapper[4914]: I1122 14:41:46.886546 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" event={"ID":"94d945b4-bca0-45e6-8c5f-c2e471e2aad5","Type":"ContainerStarted","Data":"84d05458cf1a6131eaf9d78950d3867f785d22648828c10ccad569aca52ac1ed"} Nov 22 14:41:48 crc kubenswrapper[4914]: I1122 14:41:48.904820 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" event={"ID":"94d945b4-bca0-45e6-8c5f-c2e471e2aad5","Type":"ContainerStarted","Data":"2c170d5f84f52f601e1beee573224af58e0ada252e468656c5be32f5745a47a8"} Nov 22 14:41:48 crc kubenswrapper[4914]: I1122 14:41:48.905835 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" event={"ID":"94d945b4-bca0-45e6-8c5f-c2e471e2aad5","Type":"ContainerStarted","Data":"68b21d76f1cf2eac954f39408971473136e64a724b6980f2851b757fbeaca1ea"} Nov 22 14:41:48 crc kubenswrapper[4914]: I1122 14:41:48.913439 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:49 crc kubenswrapper[4914]: I1122 14:41:49.912721 4914 generic.go:334] "Generic (PLEG): container finished" podID="713f96a8-d1f3-4fba-bd02-f025ac07a4b1" containerID="a81a6a1ee0a34f46d7875bcab79eac880ad13e2adaeb76edf7a7144f937921f7" exitCode=0 Nov 22 14:41:49 crc kubenswrapper[4914]: I1122 14:41:49.912813 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"713f96a8-d1f3-4fba-bd02-f025ac07a4b1","Type":"ContainerDied","Data":"a81a6a1ee0a34f46d7875bcab79eac880ad13e2adaeb76edf7a7144f937921f7"} Nov 22 14:41:49 crc kubenswrapper[4914]: I1122 14:41:49.948580 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" podStartSLOduration=3.354548843 podStartE2EDuration="4.948557098s" podCreationTimestamp="2025-11-22 14:41:45 +0000 UTC" firstStartedPulling="2025-11-22 14:41:46.52016035 +0000 UTC m=+872.328730031" lastFinishedPulling="2025-11-22 14:41:48.114168625 +0000 UTC m=+873.922738286" observedRunningTime="2025-11-22 14:41:48.941298113 +0000 UTC m=+874.749867794" watchObservedRunningTime="2025-11-22 14:41:49.948557098 +0000 UTC m=+875.757126759" Nov 22 14:41:50 crc kubenswrapper[4914]: I1122 14:41:50.922011 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"713f96a8-d1f3-4fba-bd02-f025ac07a4b1","Type":"ContainerStarted","Data":"9c275ac48ce97aebc999c920c2e87ec15919c89a8817d8d190675c1a1040b7b3"} Nov 22 14:41:50 crc kubenswrapper[4914]: I1122 14:41:50.924380 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:41:50 crc kubenswrapper[4914]: I1122 14:41:50.945693 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.747242214 podStartE2EDuration="41.945678105s" podCreationTimestamp="2025-11-22 14:41:09 +0000 UTC" firstStartedPulling="2025-11-22 14:41:10.898121005 +0000 UTC m=+836.706690666" lastFinishedPulling="2025-11-22 14:41:16.096556896 +0000 UTC m=+841.905126557" observedRunningTime="2025-11-22 14:41:50.941911334 +0000 UTC m=+876.750481005" watchObservedRunningTime="2025-11-22 14:41:50.945678105 +0000 UTC m=+876.754247766" Nov 22 14:41:52 crc kubenswrapper[4914]: I1122 14:41:52.042259 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:41:52 crc kubenswrapper[4914]: I1122 14:41:52.042319 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:41:56 crc kubenswrapper[4914]: I1122 14:41:56.088922 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-676b6b6b4-h8ksj" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.631642 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h"] Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.633196 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.635905 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.639161 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-4qjxm"] Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.640245 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.650704 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h"] Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.657516 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4qjxm"] Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.730382 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71cfef0-c0bc-48af-8449-3a902ce32e1f-operator-scripts\") pod \"keystone-db-create-4qjxm\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.730446 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfzt7\" (UniqueName: \"kubernetes.io/projected/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-kube-api-access-vfzt7\") pod \"keystone-0bcc-account-create-update-qpb9h\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.730490 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqggz\" (UniqueName: \"kubernetes.io/projected/d71cfef0-c0bc-48af-8449-3a902ce32e1f-kube-api-access-bqggz\") pod \"keystone-db-create-4qjxm\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.730544 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-operator-scripts\") pod \"keystone-0bcc-account-create-update-qpb9h\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.832028 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71cfef0-c0bc-48af-8449-3a902ce32e1f-operator-scripts\") pod \"keystone-db-create-4qjxm\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.832090 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfzt7\" (UniqueName: \"kubernetes.io/projected/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-kube-api-access-vfzt7\") pod \"keystone-0bcc-account-create-update-qpb9h\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.832135 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqggz\" (UniqueName: \"kubernetes.io/projected/d71cfef0-c0bc-48af-8449-3a902ce32e1f-kube-api-access-bqggz\") pod \"keystone-db-create-4qjxm\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.832188 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-operator-scripts\") pod \"keystone-0bcc-account-create-update-qpb9h\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.832931 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71cfef0-c0bc-48af-8449-3a902ce32e1f-operator-scripts\") pod \"keystone-db-create-4qjxm\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.833065 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-operator-scripts\") pod \"keystone-0bcc-account-create-update-qpb9h\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.856392 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqggz\" (UniqueName: \"kubernetes.io/projected/d71cfef0-c0bc-48af-8449-3a902ce32e1f-kube-api-access-bqggz\") pod \"keystone-db-create-4qjxm\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.863536 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfzt7\" (UniqueName: \"kubernetes.io/projected/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-kube-api-access-vfzt7\") pod \"keystone-0bcc-account-create-update-qpb9h\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.989292 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:41:59 crc kubenswrapper[4914]: I1122 14:41:59.996637 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.404071 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4qjxm"] Nov 22 14:42:00 crc kubenswrapper[4914]: W1122 14:42:00.412730 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd71cfef0_c0bc_48af_8449_3a902ce32e1f.slice/crio-afe31afa52fd0e2d1053e5d95e08710d1fd5c1d9a8cb354ceb20dd098ba3eadc WatchSource:0}: Error finding container afe31afa52fd0e2d1053e5d95e08710d1fd5c1d9a8cb354ceb20dd098ba3eadc: Status 404 returned error can't find the container with id afe31afa52fd0e2d1053e5d95e08710d1fd5c1d9a8cb354ceb20dd098ba3eadc Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.448733 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h"] Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.477542 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.994314 4914 generic.go:334] "Generic (PLEG): container finished" podID="5ba56179-4086-4862-8a37-cd9a4ccdf9bd" containerID="6a193eb20cffb18b49340e00f99afd5660b33fa54f1a570f6123fd4192399712" exitCode=0 Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.994377 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" event={"ID":"5ba56179-4086-4862-8a37-cd9a4ccdf9bd","Type":"ContainerDied","Data":"6a193eb20cffb18b49340e00f99afd5660b33fa54f1a570f6123fd4192399712"} Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.994404 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" event={"ID":"5ba56179-4086-4862-8a37-cd9a4ccdf9bd","Type":"ContainerStarted","Data":"2dcb08f6788d14e4161ce2d28db060d0a9566bc6dad9a52f8bbcb74b3d74299e"} Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.995861 4914 generic.go:334] "Generic (PLEG): container finished" podID="d71cfef0-c0bc-48af-8449-3a902ce32e1f" containerID="5b4657bfc2b93a1c1030a4f89c4f94304acf5fc4d3b2cf2f3b19bdfdedc7f9c1" exitCode=0 Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.995889 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-4qjxm" event={"ID":"d71cfef0-c0bc-48af-8449-3a902ce32e1f","Type":"ContainerDied","Data":"5b4657bfc2b93a1c1030a4f89c4f94304acf5fc4d3b2cf2f3b19bdfdedc7f9c1"} Nov 22 14:42:00 crc kubenswrapper[4914]: I1122 14:42:00.995934 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-4qjxm" event={"ID":"d71cfef0-c0bc-48af-8449-3a902ce32e1f","Type":"ContainerStarted","Data":"afe31afa52fd0e2d1053e5d95e08710d1fd5c1d9a8cb354ceb20dd098ba3eadc"} Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.319112 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-vcfbn"] Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.320380 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.322718 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-c868f" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.340435 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-vcfbn"] Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.379520 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.401105 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.469629 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqggz\" (UniqueName: \"kubernetes.io/projected/d71cfef0-c0bc-48af-8449-3a902ce32e1f-kube-api-access-bqggz\") pod \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.469717 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-operator-scripts\") pod \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.469738 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71cfef0-c0bc-48af-8449-3a902ce32e1f-operator-scripts\") pod \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\" (UID: \"d71cfef0-c0bc-48af-8449-3a902ce32e1f\") " Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.469870 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfzt7\" (UniqueName: \"kubernetes.io/projected/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-kube-api-access-vfzt7\") pod \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\" (UID: \"5ba56179-4086-4862-8a37-cd9a4ccdf9bd\") " Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.470153 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht8cw\" (UniqueName: \"kubernetes.io/projected/cd925b6d-ee1d-4af2-a157-492958a1d6ce-kube-api-access-ht8cw\") pod \"horizon-operator-index-vcfbn\" (UID: \"cd925b6d-ee1d-4af2-a157-492958a1d6ce\") " pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.470631 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71cfef0-c0bc-48af-8449-3a902ce32e1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d71cfef0-c0bc-48af-8449-3a902ce32e1f" (UID: "d71cfef0-c0bc-48af-8449-3a902ce32e1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.470700 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ba56179-4086-4862-8a37-cd9a4ccdf9bd" (UID: "5ba56179-4086-4862-8a37-cd9a4ccdf9bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.477278 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71cfef0-c0bc-48af-8449-3a902ce32e1f-kube-api-access-bqggz" (OuterVolumeSpecName: "kube-api-access-bqggz") pod "d71cfef0-c0bc-48af-8449-3a902ce32e1f" (UID: "d71cfef0-c0bc-48af-8449-3a902ce32e1f"). InnerVolumeSpecName "kube-api-access-bqggz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.477377 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-kube-api-access-vfzt7" (OuterVolumeSpecName: "kube-api-access-vfzt7") pod "5ba56179-4086-4862-8a37-cd9a4ccdf9bd" (UID: "5ba56179-4086-4862-8a37-cd9a4ccdf9bd"). InnerVolumeSpecName "kube-api-access-vfzt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.571462 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht8cw\" (UniqueName: \"kubernetes.io/projected/cd925b6d-ee1d-4af2-a157-492958a1d6ce-kube-api-access-ht8cw\") pod \"horizon-operator-index-vcfbn\" (UID: \"cd925b6d-ee1d-4af2-a157-492958a1d6ce\") " pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.571596 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfzt7\" (UniqueName: \"kubernetes.io/projected/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-kube-api-access-vfzt7\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.571608 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqggz\" (UniqueName: \"kubernetes.io/projected/d71cfef0-c0bc-48af-8449-3a902ce32e1f-kube-api-access-bqggz\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.571619 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ba56179-4086-4862-8a37-cd9a4ccdf9bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.571629 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71cfef0-c0bc-48af-8449-3a902ce32e1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.587870 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht8cw\" (UniqueName: \"kubernetes.io/projected/cd925b6d-ee1d-4af2-a157-492958a1d6ce-kube-api-access-ht8cw\") pod \"horizon-operator-index-vcfbn\" (UID: \"cd925b6d-ee1d-4af2-a157-492958a1d6ce\") " pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:02 crc kubenswrapper[4914]: I1122 14:42:02.694245 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.011281 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-4qjxm" event={"ID":"d71cfef0-c0bc-48af-8449-3a902ce32e1f","Type":"ContainerDied","Data":"afe31afa52fd0e2d1053e5d95e08710d1fd5c1d9a8cb354ceb20dd098ba3eadc"} Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.011329 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe31afa52fd0e2d1053e5d95e08710d1fd5c1d9a8cb354ceb20dd098ba3eadc" Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.011359 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4qjxm" Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.013321 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" event={"ID":"5ba56179-4086-4862-8a37-cd9a4ccdf9bd","Type":"ContainerDied","Data":"2dcb08f6788d14e4161ce2d28db060d0a9566bc6dad9a52f8bbcb74b3d74299e"} Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.013466 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dcb08f6788d14e4161ce2d28db060d0a9566bc6dad9a52f8bbcb74b3d74299e" Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.013364 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h" Nov 22 14:42:03 crc kubenswrapper[4914]: I1122 14:42:03.075683 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-vcfbn"] Nov 22 14:42:03 crc kubenswrapper[4914]: W1122 14:42:03.084348 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd925b6d_ee1d_4af2_a157_492958a1d6ce.slice/crio-174fc3696ca59d2d1bdf5d54146598b2a339aaa287356d57fbc08fcd562a53be WatchSource:0}: Error finding container 174fc3696ca59d2d1bdf5d54146598b2a339aaa287356d57fbc08fcd562a53be: Status 404 returned error can't find the container with id 174fc3696ca59d2d1bdf5d54146598b2a339aaa287356d57fbc08fcd562a53be Nov 22 14:42:04 crc kubenswrapper[4914]: I1122 14:42:04.019729 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-vcfbn" event={"ID":"cd925b6d-ee1d-4af2-a157-492958a1d6ce","Type":"ContainerStarted","Data":"174fc3696ca59d2d1bdf5d54146598b2a339aaa287356d57fbc08fcd562a53be"} Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.029353 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-vcfbn" event={"ID":"cd925b6d-ee1d-4af2-a157-492958a1d6ce","Type":"ContainerStarted","Data":"f0c67d4b8cd6d6944fa15d4906fb822b1e10bce6c2624d9b3e22e9ecfb0dc25a"} Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.048078 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-vcfbn" podStartSLOduration=2.227776918 podStartE2EDuration="3.048050419s" podCreationTimestamp="2025-11-22 14:42:02 +0000 UTC" firstStartedPulling="2025-11-22 14:42:03.086461171 +0000 UTC m=+888.895030832" lastFinishedPulling="2025-11-22 14:42:03.906734682 +0000 UTC m=+889.715304333" observedRunningTime="2025-11-22 14:42:05.044007621 +0000 UTC m=+890.852577292" watchObservedRunningTime="2025-11-22 14:42:05.048050419 +0000 UTC m=+890.856620100" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.105687 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-4xw62"] Nov 22 14:42:05 crc kubenswrapper[4914]: E1122 14:42:05.105987 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71cfef0-c0bc-48af-8449-3a902ce32e1f" containerName="mariadb-database-create" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.106002 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71cfef0-c0bc-48af-8449-3a902ce32e1f" containerName="mariadb-database-create" Nov 22 14:42:05 crc kubenswrapper[4914]: E1122 14:42:05.106017 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba56179-4086-4862-8a37-cd9a4ccdf9bd" containerName="mariadb-account-create-update" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.106025 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba56179-4086-4862-8a37-cd9a4ccdf9bd" containerName="mariadb-account-create-update" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.106169 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ba56179-4086-4862-8a37-cd9a4ccdf9bd" containerName="mariadb-account-create-update" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.106188 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71cfef0-c0bc-48af-8449-3a902ce32e1f" containerName="mariadb-database-create" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.106682 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.109207 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-tdbtf" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.123898 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-4xw62"] Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.211209 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm5kq\" (UniqueName: \"kubernetes.io/projected/b15488c7-932d-4906-8874-d71f48db7265-kube-api-access-gm5kq\") pod \"swift-operator-index-4xw62\" (UID: \"b15488c7-932d-4906-8874-d71f48db7265\") " pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.306767 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-vxptv"] Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.309102 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.311647 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.312187 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm5kq\" (UniqueName: \"kubernetes.io/projected/b15488c7-932d-4906-8874-d71f48db7265-kube-api-access-gm5kq\") pod \"swift-operator-index-4xw62\" (UID: \"b15488c7-932d-4906-8874-d71f48db7265\") " pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.312200 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-8jk5g" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.312914 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.315435 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.329421 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-vxptv"] Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.339222 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm5kq\" (UniqueName: \"kubernetes.io/projected/b15488c7-932d-4906-8874-d71f48db7265-kube-api-access-gm5kq\") pod \"swift-operator-index-4xw62\" (UID: \"b15488c7-932d-4906-8874-d71f48db7265\") " pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.413796 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzrkw\" (UniqueName: \"kubernetes.io/projected/b7c25c30-9a98-4091-be20-58ae513514bb-kube-api-access-hzrkw\") pod \"keystone-db-sync-vxptv\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.413855 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c25c30-9a98-4091-be20-58ae513514bb-config-data\") pod \"keystone-db-sync-vxptv\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.427804 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.515404 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzrkw\" (UniqueName: \"kubernetes.io/projected/b7c25c30-9a98-4091-be20-58ae513514bb-kube-api-access-hzrkw\") pod \"keystone-db-sync-vxptv\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.515700 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c25c30-9a98-4091-be20-58ae513514bb-config-data\") pod \"keystone-db-sync-vxptv\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.523017 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c25c30-9a98-4091-be20-58ae513514bb-config-data\") pod \"keystone-db-sync-vxptv\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.538566 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzrkw\" (UniqueName: \"kubernetes.io/projected/b7c25c30-9a98-4091-be20-58ae513514bb-kube-api-access-hzrkw\") pod \"keystone-db-sync-vxptv\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.630807 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:05 crc kubenswrapper[4914]: I1122 14:42:05.918053 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-4xw62"] Nov 22 14:42:05 crc kubenswrapper[4914]: W1122 14:42:05.923341 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb15488c7_932d_4906_8874_d71f48db7265.slice/crio-66faef468cc2e7b118779727727c3ced90d1ddc7acb9c29b71e2d382689e783f WatchSource:0}: Error finding container 66faef468cc2e7b118779727727c3ced90d1ddc7acb9c29b71e2d382689e783f: Status 404 returned error can't find the container with id 66faef468cc2e7b118779727727c3ced90d1ddc7acb9c29b71e2d382689e783f Nov 22 14:42:06 crc kubenswrapper[4914]: I1122 14:42:06.039185 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-4xw62" event={"ID":"b15488c7-932d-4906-8874-d71f48db7265","Type":"ContainerStarted","Data":"66faef468cc2e7b118779727727c3ced90d1ddc7acb9c29b71e2d382689e783f"} Nov 22 14:42:06 crc kubenswrapper[4914]: I1122 14:42:06.065069 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-vxptv"] Nov 22 14:42:06 crc kubenswrapper[4914]: W1122 14:42:06.068299 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7c25c30_9a98_4091_be20_58ae513514bb.slice/crio-ed90f909b690966524cb0b7c2b76633286b0159487acc9c94474c10a1262bd61 WatchSource:0}: Error finding container ed90f909b690966524cb0b7c2b76633286b0159487acc9c94474c10a1262bd61: Status 404 returned error can't find the container with id ed90f909b690966524cb0b7c2b76633286b0159487acc9c94474c10a1262bd61 Nov 22 14:42:07 crc kubenswrapper[4914]: I1122 14:42:07.048539 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-vxptv" event={"ID":"b7c25c30-9a98-4091-be20-58ae513514bb","Type":"ContainerStarted","Data":"ed90f909b690966524cb0b7c2b76633286b0159487acc9c94474c10a1262bd61"} Nov 22 14:42:12 crc kubenswrapper[4914]: I1122 14:42:12.694774 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:12 crc kubenswrapper[4914]: I1122 14:42:12.695348 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:12 crc kubenswrapper[4914]: I1122 14:42:12.721223 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:13 crc kubenswrapper[4914]: I1122 14:42:13.092792 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-vxptv" event={"ID":"b7c25c30-9a98-4091-be20-58ae513514bb","Type":"ContainerStarted","Data":"519c84eac4cdc9cf5fb89c90373219025c0bdd90711c6f90babcbe745dbfe6b4"} Nov 22 14:42:13 crc kubenswrapper[4914]: I1122 14:42:13.096183 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-4xw62" event={"ID":"b15488c7-932d-4906-8874-d71f48db7265","Type":"ContainerStarted","Data":"1b9d82f7d19b807a19cacfe1f1177ad5d78c33a6355ceaab1b5e18b1f5bc1b36"} Nov 22 14:42:13 crc kubenswrapper[4914]: I1122 14:42:13.129859 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-vxptv" podStartSLOduration=1.343958911 podStartE2EDuration="8.129831596s" podCreationTimestamp="2025-11-22 14:42:05 +0000 UTC" firstStartedPulling="2025-11-22 14:42:06.070378572 +0000 UTC m=+891.878948273" lastFinishedPulling="2025-11-22 14:42:12.856251307 +0000 UTC m=+898.664820958" observedRunningTime="2025-11-22 14:42:13.126214647 +0000 UTC m=+898.934784328" watchObservedRunningTime="2025-11-22 14:42:13.129831596 +0000 UTC m=+898.938401297" Nov 22 14:42:13 crc kubenswrapper[4914]: I1122 14:42:13.147111 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-vcfbn" Nov 22 14:42:13 crc kubenswrapper[4914]: I1122 14:42:13.149444 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-4xw62" podStartSLOduration=3.962679906 podStartE2EDuration="8.149425004s" podCreationTimestamp="2025-11-22 14:42:05 +0000 UTC" firstStartedPulling="2025-11-22 14:42:05.927661968 +0000 UTC m=+891.736231639" lastFinishedPulling="2025-11-22 14:42:10.114407066 +0000 UTC m=+895.922976737" observedRunningTime="2025-11-22 14:42:13.148587223 +0000 UTC m=+898.957156914" watchObservedRunningTime="2025-11-22 14:42:13.149425004 +0000 UTC m=+898.957994705" Nov 22 14:42:15 crc kubenswrapper[4914]: I1122 14:42:15.428262 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:15 crc kubenswrapper[4914]: I1122 14:42:15.428407 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:15 crc kubenswrapper[4914]: I1122 14:42:15.461314 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:17 crc kubenswrapper[4914]: I1122 14:42:17.127321 4914 generic.go:334] "Generic (PLEG): container finished" podID="b7c25c30-9a98-4091-be20-58ae513514bb" containerID="519c84eac4cdc9cf5fb89c90373219025c0bdd90711c6f90babcbe745dbfe6b4" exitCode=0 Nov 22 14:42:17 crc kubenswrapper[4914]: I1122 14:42:17.127450 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-vxptv" event={"ID":"b7c25c30-9a98-4091-be20-58ae513514bb","Type":"ContainerDied","Data":"519c84eac4cdc9cf5fb89c90373219025c0bdd90711c6f90babcbe745dbfe6b4"} Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.495155 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.599308 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c25c30-9a98-4091-be20-58ae513514bb-config-data\") pod \"b7c25c30-9a98-4091-be20-58ae513514bb\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.599386 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzrkw\" (UniqueName: \"kubernetes.io/projected/b7c25c30-9a98-4091-be20-58ae513514bb-kube-api-access-hzrkw\") pod \"b7c25c30-9a98-4091-be20-58ae513514bb\" (UID: \"b7c25c30-9a98-4091-be20-58ae513514bb\") " Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.605721 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7c25c30-9a98-4091-be20-58ae513514bb-kube-api-access-hzrkw" (OuterVolumeSpecName: "kube-api-access-hzrkw") pod "b7c25c30-9a98-4091-be20-58ae513514bb" (UID: "b7c25c30-9a98-4091-be20-58ae513514bb"). InnerVolumeSpecName "kube-api-access-hzrkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.649447 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7c25c30-9a98-4091-be20-58ae513514bb-config-data" (OuterVolumeSpecName: "config-data") pod "b7c25c30-9a98-4091-be20-58ae513514bb" (UID: "b7c25c30-9a98-4091-be20-58ae513514bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.701892 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7c25c30-9a98-4091-be20-58ae513514bb-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:18 crc kubenswrapper[4914]: I1122 14:42:18.701935 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzrkw\" (UniqueName: \"kubernetes.io/projected/b7c25c30-9a98-4091-be20-58ae513514bb-kube-api-access-hzrkw\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.145978 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-vxptv" event={"ID":"b7c25c30-9a98-4091-be20-58ae513514bb","Type":"ContainerDied","Data":"ed90f909b690966524cb0b7c2b76633286b0159487acc9c94474c10a1262bd61"} Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.146031 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed90f909b690966524cb0b7c2b76633286b0159487acc9c94474c10a1262bd61" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.146054 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-vxptv" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.347044 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-px7jm"] Nov 22 14:42:19 crc kubenswrapper[4914]: E1122 14:42:19.347705 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c25c30-9a98-4091-be20-58ae513514bb" containerName="keystone-db-sync" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.347748 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c25c30-9a98-4091-be20-58ae513514bb" containerName="keystone-db-sync" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.348046 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7c25c30-9a98-4091-be20-58ae513514bb" containerName="keystone-db-sync" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.350056 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.356347 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.356426 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.356360 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-8jk5g" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.357603 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.359205 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.359997 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-px7jm"] Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.413170 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-fernet-keys\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.413242 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st7j5\" (UniqueName: \"kubernetes.io/projected/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-kube-api-access-st7j5\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.413280 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-credential-keys\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.413385 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-scripts\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.413415 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-config-data\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.514476 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-scripts\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.514521 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-config-data\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.514564 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-fernet-keys\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.514630 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st7j5\" (UniqueName: \"kubernetes.io/projected/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-kube-api-access-st7j5\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.514658 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-credential-keys\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.518038 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-fernet-keys\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.518397 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-scripts\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.519607 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-config-data\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.519796 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-credential-keys\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.531682 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st7j5\" (UniqueName: \"kubernetes.io/projected/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-kube-api-access-st7j5\") pod \"keystone-bootstrap-px7jm\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.673355 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:19 crc kubenswrapper[4914]: I1122 14:42:19.954172 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-px7jm"] Nov 22 14:42:20 crc kubenswrapper[4914]: I1122 14:42:20.152812 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" event={"ID":"5b65bb30-a8e0-44b3-8400-9bcd7d99a184","Type":"ContainerStarted","Data":"bcea9dad0d1a61b1b0b50d60065b2a16a7bba31f6eea656199b42b259311e8ca"} Nov 22 14:42:20 crc kubenswrapper[4914]: I1122 14:42:20.153174 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" event={"ID":"5b65bb30-a8e0-44b3-8400-9bcd7d99a184","Type":"ContainerStarted","Data":"03a5b6d899ef4319905e9f760a497a6f23ed19d18a3f312fcf37b806807de4a2"} Nov 22 14:42:20 crc kubenswrapper[4914]: I1122 14:42:20.174186 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" podStartSLOduration=1.174165879 podStartE2EDuration="1.174165879s" podCreationTimestamp="2025-11-22 14:42:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:42:20.170718555 +0000 UTC m=+905.979288216" watchObservedRunningTime="2025-11-22 14:42:20.174165879 +0000 UTC m=+905.982735540" Nov 22 14:42:22 crc kubenswrapper[4914]: I1122 14:42:22.043185 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:42:22 crc kubenswrapper[4914]: I1122 14:42:22.043692 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:42:23 crc kubenswrapper[4914]: I1122 14:42:23.181500 4914 generic.go:334] "Generic (PLEG): container finished" podID="5b65bb30-a8e0-44b3-8400-9bcd7d99a184" containerID="bcea9dad0d1a61b1b0b50d60065b2a16a7bba31f6eea656199b42b259311e8ca" exitCode=0 Nov 22 14:42:23 crc kubenswrapper[4914]: I1122 14:42:23.181552 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" event={"ID":"5b65bb30-a8e0-44b3-8400-9bcd7d99a184","Type":"ContainerDied","Data":"bcea9dad0d1a61b1b0b50d60065b2a16a7bba31f6eea656199b42b259311e8ca"} Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.548949 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.597634 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-credential-keys\") pod \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.597721 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st7j5\" (UniqueName: \"kubernetes.io/projected/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-kube-api-access-st7j5\") pod \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.597771 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-fernet-keys\") pod \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.597825 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-scripts\") pod \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.597868 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-config-data\") pod \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\" (UID: \"5b65bb30-a8e0-44b3-8400-9bcd7d99a184\") " Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.606126 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5b65bb30-a8e0-44b3-8400-9bcd7d99a184" (UID: "5b65bb30-a8e0-44b3-8400-9bcd7d99a184"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.607470 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-kube-api-access-st7j5" (OuterVolumeSpecName: "kube-api-access-st7j5") pod "5b65bb30-a8e0-44b3-8400-9bcd7d99a184" (UID: "5b65bb30-a8e0-44b3-8400-9bcd7d99a184"). InnerVolumeSpecName "kube-api-access-st7j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.614304 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-scripts" (OuterVolumeSpecName: "scripts") pod "5b65bb30-a8e0-44b3-8400-9bcd7d99a184" (UID: "5b65bb30-a8e0-44b3-8400-9bcd7d99a184"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.614494 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5b65bb30-a8e0-44b3-8400-9bcd7d99a184" (UID: "5b65bb30-a8e0-44b3-8400-9bcd7d99a184"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.625678 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-config-data" (OuterVolumeSpecName: "config-data") pod "5b65bb30-a8e0-44b3-8400-9bcd7d99a184" (UID: "5b65bb30-a8e0-44b3-8400-9bcd7d99a184"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.701377 4914 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.701411 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st7j5\" (UniqueName: \"kubernetes.io/projected/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-kube-api-access-st7j5\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.701429 4914 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.701440 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:24 crc kubenswrapper[4914]: I1122 14:42:24.701451 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b65bb30-a8e0-44b3-8400-9bcd7d99a184-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.203390 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" event={"ID":"5b65bb30-a8e0-44b3-8400-9bcd7d99a184","Type":"ContainerDied","Data":"03a5b6d899ef4319905e9f760a497a6f23ed19d18a3f312fcf37b806807de4a2"} Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.203450 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03a5b6d899ef4319905e9f760a497a6f23ed19d18a3f312fcf37b806807de4a2" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.203489 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-px7jm" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.320851 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-9c89d49d9-jhwn4"] Nov 22 14:42:25 crc kubenswrapper[4914]: E1122 14:42:25.321350 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b65bb30-a8e0-44b3-8400-9bcd7d99a184" containerName="keystone-bootstrap" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.321395 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b65bb30-a8e0-44b3-8400-9bcd7d99a184" containerName="keystone-bootstrap" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.321732 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b65bb30-a8e0-44b3-8400-9bcd7d99a184" containerName="keystone-bootstrap" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.323668 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.326624 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.326727 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.327019 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.330288 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-8jk5g" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.333191 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-9c89d49d9-jhwn4"] Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.411312 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-credential-keys\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.411482 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-fernet-keys\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.411556 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-config-data\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.411636 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-scripts\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.411678 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqsht\" (UniqueName: \"kubernetes.io/projected/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-kube-api-access-tqsht\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.458089 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-4xw62" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.512651 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-scripts\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.512718 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqsht\" (UniqueName: \"kubernetes.io/projected/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-kube-api-access-tqsht\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.512808 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-credential-keys\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.512864 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-fernet-keys\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.512937 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-config-data\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.518074 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-config-data\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.519079 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-scripts\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.519518 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-credential-keys\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.521705 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-fernet-keys\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.559859 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqsht\" (UniqueName: \"kubernetes.io/projected/1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f-kube-api-access-tqsht\") pod \"keystone-9c89d49d9-jhwn4\" (UID: \"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f\") " pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:25 crc kubenswrapper[4914]: I1122 14:42:25.652194 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:26 crc kubenswrapper[4914]: I1122 14:42:26.043352 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-9c89d49d9-jhwn4"] Nov 22 14:42:26 crc kubenswrapper[4914]: I1122 14:42:26.215298 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" event={"ID":"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f","Type":"ContainerStarted","Data":"f0670095c42c6dd77f95705862f54cdf214873e1487934b083ea08504fbdb137"} Nov 22 14:42:26 crc kubenswrapper[4914]: I1122 14:42:26.215368 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" event={"ID":"1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f","Type":"ContainerStarted","Data":"167bcb7ed4d93affb0cc5a6a756df54f62b6f25cd266e3350ceb249dcac7c8f4"} Nov 22 14:42:26 crc kubenswrapper[4914]: I1122 14:42:26.215431 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:26 crc kubenswrapper[4914]: I1122 14:42:26.247528 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" podStartSLOduration=1.247509803 podStartE2EDuration="1.247509803s" podCreationTimestamp="2025-11-22 14:42:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:42:26.24365307 +0000 UTC m=+912.052222801" watchObservedRunningTime="2025-11-22 14:42:26.247509803 +0000 UTC m=+912.056079464" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.373198 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5"] Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.377536 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.380297 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l5bmb" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.405023 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5"] Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.577375 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-bundle\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.577434 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-util\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.577507 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6r9z\" (UniqueName: \"kubernetes.io/projected/d95bcff7-8ac6-44ee-84b1-83c955faa3af-kube-api-access-n6r9z\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.679175 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-bundle\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.679245 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-util\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.679315 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6r9z\" (UniqueName: \"kubernetes.io/projected/d95bcff7-8ac6-44ee-84b1-83c955faa3af-kube-api-access-n6r9z\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.679731 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-bundle\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.679954 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-util\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:35 crc kubenswrapper[4914]: I1122 14:42:35.704466 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6r9z\" (UniqueName: \"kubernetes.io/projected/d95bcff7-8ac6-44ee-84b1-83c955faa3af-kube-api-access-n6r9z\") pod \"440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.003482 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.318921 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5"] Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.369629 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p"] Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.371080 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.390849 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p"] Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.393164 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-bundle\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.393223 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-util\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.393349 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghpbk\" (UniqueName: \"kubernetes.io/projected/8d671252-1412-4afd-8c99-ac3540bb8430-kube-api-access-ghpbk\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.495134 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghpbk\" (UniqueName: \"kubernetes.io/projected/8d671252-1412-4afd-8c99-ac3540bb8430-kube-api-access-ghpbk\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.495461 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-bundle\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.495493 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-util\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.495935 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-bundle\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.496013 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-util\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.518845 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghpbk\" (UniqueName: \"kubernetes.io/projected/8d671252-1412-4afd-8c99-ac3540bb8430-kube-api-access-ghpbk\") pod \"62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:36 crc kubenswrapper[4914]: I1122 14:42:36.710303 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:37 crc kubenswrapper[4914]: I1122 14:42:37.188908 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p"] Nov 22 14:42:37 crc kubenswrapper[4914]: I1122 14:42:37.318766 4914 generic.go:334] "Generic (PLEG): container finished" podID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerID="1064ebfbba5d54c7c4627bb9db52b721e2821c66dc9e65e9ad132c58b4a53d48" exitCode=0 Nov 22 14:42:37 crc kubenswrapper[4914]: I1122 14:42:37.318929 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" event={"ID":"d95bcff7-8ac6-44ee-84b1-83c955faa3af","Type":"ContainerDied","Data":"1064ebfbba5d54c7c4627bb9db52b721e2821c66dc9e65e9ad132c58b4a53d48"} Nov 22 14:42:37 crc kubenswrapper[4914]: I1122 14:42:37.318971 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" event={"ID":"d95bcff7-8ac6-44ee-84b1-83c955faa3af","Type":"ContainerStarted","Data":"594e4f46152e1a5190e39276d07c4c3d859d512338716495a800b208d84ad33d"} Nov 22 14:42:37 crc kubenswrapper[4914]: I1122 14:42:37.321728 4914 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 14:42:37 crc kubenswrapper[4914]: I1122 14:42:37.321778 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" event={"ID":"8d671252-1412-4afd-8c99-ac3540bb8430","Type":"ContainerStarted","Data":"1d3a5a3ddd4f80b6c4c7a4e503584e0b575ccadbecd9bb409c16bc01e7320fe5"} Nov 22 14:42:38 crc kubenswrapper[4914]: I1122 14:42:38.333123 4914 generic.go:334] "Generic (PLEG): container finished" podID="8d671252-1412-4afd-8c99-ac3540bb8430" containerID="8dccfcc69ad3919e50f93a7315c6f683ab4a5b7fe8e3401d40a5d133ee959008" exitCode=0 Nov 22 14:42:38 crc kubenswrapper[4914]: I1122 14:42:38.333173 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" event={"ID":"8d671252-1412-4afd-8c99-ac3540bb8430","Type":"ContainerDied","Data":"8dccfcc69ad3919e50f93a7315c6f683ab4a5b7fe8e3401d40a5d133ee959008"} Nov 22 14:42:40 crc kubenswrapper[4914]: I1122 14:42:40.352324 4914 generic.go:334] "Generic (PLEG): container finished" podID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerID="c1206f667470c6fcd81ab7f72415573385ed6a1fd8b17d6687f6518178436610" exitCode=0 Nov 22 14:42:40 crc kubenswrapper[4914]: I1122 14:42:40.352504 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" event={"ID":"d95bcff7-8ac6-44ee-84b1-83c955faa3af","Type":"ContainerDied","Data":"c1206f667470c6fcd81ab7f72415573385ed6a1fd8b17d6687f6518178436610"} Nov 22 14:42:41 crc kubenswrapper[4914]: I1122 14:42:41.387818 4914 generic.go:334] "Generic (PLEG): container finished" podID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerID="12ef50edf177c1427144ff45504a971106c7ef041ebc059e34b9839dcf7cbdfb" exitCode=0 Nov 22 14:42:41 crc kubenswrapper[4914]: I1122 14:42:41.387940 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" event={"ID":"d95bcff7-8ac6-44ee-84b1-83c955faa3af","Type":"ContainerDied","Data":"12ef50edf177c1427144ff45504a971106c7ef041ebc059e34b9839dcf7cbdfb"} Nov 22 14:42:41 crc kubenswrapper[4914]: I1122 14:42:41.390442 4914 generic.go:334] "Generic (PLEG): container finished" podID="8d671252-1412-4afd-8c99-ac3540bb8430" containerID="1facb04f23c81f2e9ba9ccbb1e7d799ce7f12b8dda989f66cc8de707b478838d" exitCode=0 Nov 22 14:42:41 crc kubenswrapper[4914]: I1122 14:42:41.390502 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" event={"ID":"8d671252-1412-4afd-8c99-ac3540bb8430","Type":"ContainerDied","Data":"1facb04f23c81f2e9ba9ccbb1e7d799ce7f12b8dda989f66cc8de707b478838d"} Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.400559 4914 generic.go:334] "Generic (PLEG): container finished" podID="8d671252-1412-4afd-8c99-ac3540bb8430" containerID="658defbb43f040a04aa1084dd26564b63acbfec52437d9ccddd3bd980464a26f" exitCode=0 Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.400630 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" event={"ID":"8d671252-1412-4afd-8c99-ac3540bb8430","Type":"ContainerDied","Data":"658defbb43f040a04aa1084dd26564b63acbfec52437d9ccddd3bd980464a26f"} Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.761844 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.796020 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-util\") pod \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.796143 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6r9z\" (UniqueName: \"kubernetes.io/projected/d95bcff7-8ac6-44ee-84b1-83c955faa3af-kube-api-access-n6r9z\") pod \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.796176 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-bundle\") pod \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\" (UID: \"d95bcff7-8ac6-44ee-84b1-83c955faa3af\") " Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.797029 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-bundle" (OuterVolumeSpecName: "bundle") pod "d95bcff7-8ac6-44ee-84b1-83c955faa3af" (UID: "d95bcff7-8ac6-44ee-84b1-83c955faa3af"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.803011 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95bcff7-8ac6-44ee-84b1-83c955faa3af-kube-api-access-n6r9z" (OuterVolumeSpecName: "kube-api-access-n6r9z") pod "d95bcff7-8ac6-44ee-84b1-83c955faa3af" (UID: "d95bcff7-8ac6-44ee-84b1-83c955faa3af"). InnerVolumeSpecName "kube-api-access-n6r9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.816027 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-util" (OuterVolumeSpecName: "util") pod "d95bcff7-8ac6-44ee-84b1-83c955faa3af" (UID: "d95bcff7-8ac6-44ee-84b1-83c955faa3af"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.898526 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.898567 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6r9z\" (UniqueName: \"kubernetes.io/projected/d95bcff7-8ac6-44ee-84b1-83c955faa3af-kube-api-access-n6r9z\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:42 crc kubenswrapper[4914]: I1122 14:42:42.898580 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d95bcff7-8ac6-44ee-84b1-83c955faa3af-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.409454 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" event={"ID":"d95bcff7-8ac6-44ee-84b1-83c955faa3af","Type":"ContainerDied","Data":"594e4f46152e1a5190e39276d07c4c3d859d512338716495a800b208d84ad33d"} Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.409503 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="594e4f46152e1a5190e39276d07c4c3d859d512338716495a800b208d84ad33d" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.409472 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.710861 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.814177 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-util\") pod \"8d671252-1412-4afd-8c99-ac3540bb8430\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.814314 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-bundle\") pod \"8d671252-1412-4afd-8c99-ac3540bb8430\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.814391 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghpbk\" (UniqueName: \"kubernetes.io/projected/8d671252-1412-4afd-8c99-ac3540bb8430-kube-api-access-ghpbk\") pod \"8d671252-1412-4afd-8c99-ac3540bb8430\" (UID: \"8d671252-1412-4afd-8c99-ac3540bb8430\") " Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.815507 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-bundle" (OuterVolumeSpecName: "bundle") pod "8d671252-1412-4afd-8c99-ac3540bb8430" (UID: "8d671252-1412-4afd-8c99-ac3540bb8430"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.820377 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d671252-1412-4afd-8c99-ac3540bb8430-kube-api-access-ghpbk" (OuterVolumeSpecName: "kube-api-access-ghpbk") pod "8d671252-1412-4afd-8c99-ac3540bb8430" (UID: "8d671252-1412-4afd-8c99-ac3540bb8430"). InnerVolumeSpecName "kube-api-access-ghpbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.827346 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-util" (OuterVolumeSpecName: "util") pod "8d671252-1412-4afd-8c99-ac3540bb8430" (UID: "8d671252-1412-4afd-8c99-ac3540bb8430"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.916773 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.916812 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghpbk\" (UniqueName: \"kubernetes.io/projected/8d671252-1412-4afd-8c99-ac3540bb8430-kube-api-access-ghpbk\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:43 crc kubenswrapper[4914]: I1122 14:42:43.916828 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8d671252-1412-4afd-8c99-ac3540bb8430-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:42:44 crc kubenswrapper[4914]: I1122 14:42:44.422331 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" event={"ID":"8d671252-1412-4afd-8c99-ac3540bb8430","Type":"ContainerDied","Data":"1d3a5a3ddd4f80b6c4c7a4e503584e0b575ccadbecd9bb409c16bc01e7320fe5"} Nov 22 14:42:44 crc kubenswrapper[4914]: I1122 14:42:44.422769 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d3a5a3ddd4f80b6c4c7a4e503584e0b575ccadbecd9bb409c16bc01e7320fe5" Nov 22 14:42:44 crc kubenswrapper[4914]: I1122 14:42:44.422503 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p" Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.042718 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.043724 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.043811 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.044843 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa8dd9653a1cc07eb97e4563bba7df9b08941e911b0fd3c2ed65154f5dab470f"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.044999 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://aa8dd9653a1cc07eb97e4563bba7df9b08941e911b0fd3c2ed65154f5dab470f" gracePeriod=600 Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.492816 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="aa8dd9653a1cc07eb97e4563bba7df9b08941e911b0fd3c2ed65154f5dab470f" exitCode=0 Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.492909 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"aa8dd9653a1cc07eb97e4563bba7df9b08941e911b0fd3c2ed65154f5dab470f"} Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.493247 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"d84b3eee4b009a037e512f975ccca21eff513d79b47374207f1e9f4353ec16da"} Nov 22 14:42:52 crc kubenswrapper[4914]: I1122 14:42:52.493275 4914 scope.go:117] "RemoveContainer" containerID="448c460df7aa9c496e7a50b996cc4b57b27198da3e1c4c2cbdb4f3896aa006fa" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.070533 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz"] Nov 22 14:42:54 crc kubenswrapper[4914]: E1122 14:42:54.072038 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="pull" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.072144 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="pull" Nov 22 14:42:54 crc kubenswrapper[4914]: E1122 14:42:54.072225 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="util" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.072308 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="util" Nov 22 14:42:54 crc kubenswrapper[4914]: E1122 14:42:54.072395 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="extract" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.072464 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="extract" Nov 22 14:42:54 crc kubenswrapper[4914]: E1122 14:42:54.072548 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="util" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.072635 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="util" Nov 22 14:42:54 crc kubenswrapper[4914]: E1122 14:42:54.072724 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="extract" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.072820 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="extract" Nov 22 14:42:54 crc kubenswrapper[4914]: E1122 14:42:54.072913 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="pull" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.073001 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="pull" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.073204 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d671252-1412-4afd-8c99-ac3540bb8430" containerName="extract" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.073305 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95bcff7-8ac6-44ee-84b1-83c955faa3af" containerName="extract" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.074118 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.077005 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bw2pt" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.077180 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.083867 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz"] Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.182804 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8699b0cd-d651-448d-9750-812ef636a23f-webhook-cert\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.183183 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c462c\" (UniqueName: \"kubernetes.io/projected/8699b0cd-d651-448d-9750-812ef636a23f-kube-api-access-c462c\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.183222 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8699b0cd-d651-448d-9750-812ef636a23f-apiservice-cert\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.284390 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8699b0cd-d651-448d-9750-812ef636a23f-apiservice-cert\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.284510 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8699b0cd-d651-448d-9750-812ef636a23f-webhook-cert\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.284585 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c462c\" (UniqueName: \"kubernetes.io/projected/8699b0cd-d651-448d-9750-812ef636a23f-kube-api-access-c462c\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.296470 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8699b0cd-d651-448d-9750-812ef636a23f-webhook-cert\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.304186 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8699b0cd-d651-448d-9750-812ef636a23f-apiservice-cert\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.310667 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c462c\" (UniqueName: \"kubernetes.io/projected/8699b0cd-d651-448d-9750-812ef636a23f-kube-api-access-c462c\") pod \"swift-operator-controller-manager-8497df8d9f-s9nlz\" (UID: \"8699b0cd-d651-448d-9750-812ef636a23f\") " pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.393607 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:54 crc kubenswrapper[4914]: I1122 14:42:54.816600 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz"] Nov 22 14:42:55 crc kubenswrapper[4914]: I1122 14:42:55.519451 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" event={"ID":"8699b0cd-d651-448d-9750-812ef636a23f","Type":"ContainerStarted","Data":"cb60419611c90e150870a960959c0e55d16d97807fd745dba98b6892e8ec980e"} Nov 22 14:42:57 crc kubenswrapper[4914]: I1122 14:42:57.020365 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-9c89d49d9-jhwn4" Nov 22 14:42:57 crc kubenswrapper[4914]: I1122 14:42:57.561830 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" event={"ID":"8699b0cd-d651-448d-9750-812ef636a23f","Type":"ContainerStarted","Data":"586e29d6319c5939e8fa307b4abf4990f5910d2189eef53d6df4bbb48d9b277b"} Nov 22 14:42:57 crc kubenswrapper[4914]: I1122 14:42:57.561891 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" event={"ID":"8699b0cd-d651-448d-9750-812ef636a23f","Type":"ContainerStarted","Data":"c2c1e3a6e7d2b1bc9117c6193dd61f5fe9113a2f29ac10b95c28cb8df2be2d69"} Nov 22 14:42:57 crc kubenswrapper[4914]: I1122 14:42:57.562078 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:42:57 crc kubenswrapper[4914]: I1122 14:42:57.587684 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" podStartSLOduration=1.9580692640000001 podStartE2EDuration="3.587663037s" podCreationTimestamp="2025-11-22 14:42:54 +0000 UTC" firstStartedPulling="2025-11-22 14:42:54.834031129 +0000 UTC m=+940.642600790" lastFinishedPulling="2025-11-22 14:42:56.463624902 +0000 UTC m=+942.272194563" observedRunningTime="2025-11-22 14:42:57.585366241 +0000 UTC m=+943.393935902" watchObservedRunningTime="2025-11-22 14:42:57.587663037 +0000 UTC m=+943.396232698" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.498357 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt"] Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.500087 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.506267 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-b5r99" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.511981 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt"] Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.514597 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.693599 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-apiservice-cert\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.693641 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-webhook-cert\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.693784 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jr52\" (UniqueName: \"kubernetes.io/projected/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-kube-api-access-5jr52\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.795485 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-apiservice-cert\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.795537 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-webhook-cert\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.795572 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jr52\" (UniqueName: \"kubernetes.io/projected/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-kube-api-access-5jr52\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.801378 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-webhook-cert\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.802690 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-apiservice-cert\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.815503 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jr52\" (UniqueName: \"kubernetes.io/projected/18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a-kube-api-access-5jr52\") pod \"horizon-operator-controller-manager-7f7479bbf4-hbvtt\" (UID: \"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a\") " pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:00 crc kubenswrapper[4914]: I1122 14:43:00.820541 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:01 crc kubenswrapper[4914]: I1122 14:43:01.045950 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt"] Nov 22 14:43:01 crc kubenswrapper[4914]: I1122 14:43:01.593825 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" event={"ID":"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a","Type":"ContainerStarted","Data":"d90141109ff38ad9ad1636cd976eee77c5682584865a5a5cd3906f8ae04eff2c"} Nov 22 14:43:03 crc kubenswrapper[4914]: I1122 14:43:03.610170 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" event={"ID":"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a","Type":"ContainerStarted","Data":"b1080dc5eaf1686f3cfb8153fa04dcea577d3ab5425a040728b7545877abf982"} Nov 22 14:43:03 crc kubenswrapper[4914]: I1122 14:43:03.610734 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:03 crc kubenswrapper[4914]: I1122 14:43:03.610745 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" event={"ID":"18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a","Type":"ContainerStarted","Data":"36c68455b7348e1b60f035bb88221ae2d26219f8cb5c7fd3c956976ceefa4990"} Nov 22 14:43:03 crc kubenswrapper[4914]: I1122 14:43:03.631042 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" podStartSLOduration=1.998855193 podStartE2EDuration="3.63101977s" podCreationTimestamp="2025-11-22 14:43:00 +0000 UTC" firstStartedPulling="2025-11-22 14:43:01.053253333 +0000 UTC m=+946.861822984" lastFinishedPulling="2025-11-22 14:43:02.6854179 +0000 UTC m=+948.493987561" observedRunningTime="2025-11-22 14:43:03.6248676 +0000 UTC m=+949.433437291" watchObservedRunningTime="2025-11-22 14:43:03.63101977 +0000 UTC m=+949.439589441" Nov 22 14:43:04 crc kubenswrapper[4914]: I1122 14:43:04.407957 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-8497df8d9f-s9nlz" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.172510 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.176559 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.179754 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-xdprc" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.179780 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.180019 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.188123 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.207270 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.289665 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cfa8c85b-6b20-4bde-a701-06b713f81c06-lock\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.289716 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cfa8c85b-6b20-4bde-a701-06b713f81c06-cache\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.289740 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sxdm\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-kube-api-access-6sxdm\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.289987 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.290073 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.391520 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cfa8c85b-6b20-4bde-a701-06b713f81c06-lock\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.391567 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cfa8c85b-6b20-4bde-a701-06b713f81c06-cache\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.391590 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sxdm\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-kube-api-access-6sxdm\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.391630 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.391655 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: E1122 14:43:07.391782 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:07 crc kubenswrapper[4914]: E1122 14:43:07.391795 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 22 14:43:07 crc kubenswrapper[4914]: E1122 14:43:07.391870 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift podName:cfa8c85b-6b20-4bde-a701-06b713f81c06 nodeName:}" failed. No retries permitted until 2025-11-22 14:43:07.891855722 +0000 UTC m=+953.700425383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift") pod "swift-storage-0" (UID: "cfa8c85b-6b20-4bde-a701-06b713f81c06") : configmap "swift-ring-files" not found Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.392212 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.392740 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cfa8c85b-6b20-4bde-a701-06b713f81c06-lock\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.392819 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cfa8c85b-6b20-4bde-a701-06b713f81c06-cache\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.417864 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sxdm\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-kube-api-access-6sxdm\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.430784 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.897534 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:07 crc kubenswrapper[4914]: E1122 14:43:07.897769 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:07 crc kubenswrapper[4914]: E1122 14:43:07.897782 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 22 14:43:07 crc kubenswrapper[4914]: E1122 14:43:07.897830 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift podName:cfa8c85b-6b20-4bde-a701-06b713f81c06 nodeName:}" failed. No retries permitted until 2025-11-22 14:43:08.897815062 +0000 UTC m=+954.706384723 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift") pod "swift-storage-0" (UID: "cfa8c85b-6b20-4bde-a701-06b713f81c06") : configmap "swift-ring-files" not found Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.916321 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-bggm4"] Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.917313 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.919548 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-h5zpx" Nov 22 14:43:07 crc kubenswrapper[4914]: I1122 14:43:07.930611 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-bggm4"] Nov 22 14:43:08 crc kubenswrapper[4914]: I1122 14:43:08.000511 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9st4v\" (UniqueName: \"kubernetes.io/projected/293f2650-2180-4730-9572-c5c59e370c97-kube-api-access-9st4v\") pod \"glance-operator-index-bggm4\" (UID: \"293f2650-2180-4730-9572-c5c59e370c97\") " pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:08 crc kubenswrapper[4914]: I1122 14:43:08.101980 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9st4v\" (UniqueName: \"kubernetes.io/projected/293f2650-2180-4730-9572-c5c59e370c97-kube-api-access-9st4v\") pod \"glance-operator-index-bggm4\" (UID: \"293f2650-2180-4730-9572-c5c59e370c97\") " pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:08 crc kubenswrapper[4914]: I1122 14:43:08.119505 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9st4v\" (UniqueName: \"kubernetes.io/projected/293f2650-2180-4730-9572-c5c59e370c97-kube-api-access-9st4v\") pod \"glance-operator-index-bggm4\" (UID: \"293f2650-2180-4730-9572-c5c59e370c97\") " pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:08 crc kubenswrapper[4914]: I1122 14:43:08.242535 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:08 crc kubenswrapper[4914]: I1122 14:43:08.681773 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-bggm4"] Nov 22 14:43:08 crc kubenswrapper[4914]: W1122 14:43:08.683045 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod293f2650_2180_4730_9572_c5c59e370c97.slice/crio-3a32f2f38b508313d276eec1588bc5348380b018b45404bbcf4bfef86309f1b7 WatchSource:0}: Error finding container 3a32f2f38b508313d276eec1588bc5348380b018b45404bbcf4bfef86309f1b7: Status 404 returned error can't find the container with id 3a32f2f38b508313d276eec1588bc5348380b018b45404bbcf4bfef86309f1b7 Nov 22 14:43:08 crc kubenswrapper[4914]: I1122 14:43:08.914046 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:08 crc kubenswrapper[4914]: E1122 14:43:08.914288 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:08 crc kubenswrapper[4914]: E1122 14:43:08.914764 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 22 14:43:08 crc kubenswrapper[4914]: E1122 14:43:08.914891 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift podName:cfa8c85b-6b20-4bde-a701-06b713f81c06 nodeName:}" failed. No retries permitted until 2025-11-22 14:43:10.914862905 +0000 UTC m=+956.723432566 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift") pod "swift-storage-0" (UID: "cfa8c85b-6b20-4bde-a701-06b713f81c06") : configmap "swift-ring-files" not found Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.565313 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-547856594f-9j8pz"] Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.566511 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.570616 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.592189 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-547856594f-9j8pz"] Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.627428 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779691-44d2-40cd-8138-03953251ba7c-config-data\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.627488 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76mkw\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-kube-api-access-76mkw\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.627515 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7779691-44d2-40cd-8138-03953251ba7c-log-httpd\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.627544 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7779691-44d2-40cd-8138-03953251ba7c-run-httpd\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.627560 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.656162 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bggm4" event={"ID":"293f2650-2180-4730-9572-c5c59e370c97","Type":"ContainerStarted","Data":"3a32f2f38b508313d276eec1588bc5348380b018b45404bbcf4bfef86309f1b7"} Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.728909 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779691-44d2-40cd-8138-03953251ba7c-config-data\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.728965 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76mkw\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-kube-api-access-76mkw\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.728986 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7779691-44d2-40cd-8138-03953251ba7c-log-httpd\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.729012 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7779691-44d2-40cd-8138-03953251ba7c-run-httpd\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.729030 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.729579 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7779691-44d2-40cd-8138-03953251ba7c-run-httpd\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: E1122 14:43:09.729697 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:09 crc kubenswrapper[4914]: E1122 14:43:09.729771 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-547856594f-9j8pz: configmap "swift-ring-files" not found Nov 22 14:43:09 crc kubenswrapper[4914]: E1122 14:43:09.729828 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift podName:f7779691-44d2-40cd-8138-03953251ba7c nodeName:}" failed. No retries permitted until 2025-11-22 14:43:10.229811746 +0000 UTC m=+956.038381397 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift") pod "swift-proxy-547856594f-9j8pz" (UID: "f7779691-44d2-40cd-8138-03953251ba7c") : configmap "swift-ring-files" not found Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.731128 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7779691-44d2-40cd-8138-03953251ba7c-log-httpd\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.739145 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7779691-44d2-40cd-8138-03953251ba7c-config-data\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:09 crc kubenswrapper[4914]: I1122 14:43:09.754294 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76mkw\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-kube-api-access-76mkw\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:10 crc kubenswrapper[4914]: I1122 14:43:10.238357 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:10 crc kubenswrapper[4914]: E1122 14:43:10.238489 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:10 crc kubenswrapper[4914]: E1122 14:43:10.238502 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-547856594f-9j8pz: configmap "swift-ring-files" not found Nov 22 14:43:10 crc kubenswrapper[4914]: E1122 14:43:10.238543 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift podName:f7779691-44d2-40cd-8138-03953251ba7c nodeName:}" failed. No retries permitted until 2025-11-22 14:43:11.238528712 +0000 UTC m=+957.047098373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift") pod "swift-proxy-547856594f-9j8pz" (UID: "f7779691-44d2-40cd-8138-03953251ba7c") : configmap "swift-ring-files" not found Nov 22 14:43:10 crc kubenswrapper[4914]: I1122 14:43:10.824623 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7f7479bbf4-hbvtt" Nov 22 14:43:10 crc kubenswrapper[4914]: I1122 14:43:10.947781 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:10 crc kubenswrapper[4914]: E1122 14:43:10.947992 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:10 crc kubenswrapper[4914]: E1122 14:43:10.948008 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 22 14:43:10 crc kubenswrapper[4914]: E1122 14:43:10.948064 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift podName:cfa8c85b-6b20-4bde-a701-06b713f81c06 nodeName:}" failed. No retries permitted until 2025-11-22 14:43:14.94804429 +0000 UTC m=+960.756613951 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift") pod "swift-storage-0" (UID: "cfa8c85b-6b20-4bde-a701-06b713f81c06") : configmap "swift-ring-files" not found Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.224231 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-zfxrw"] Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.225263 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.227473 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.227647 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.235925 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-zfxrw"] Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.252355 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:11 crc kubenswrapper[4914]: E1122 14:43:11.252547 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:11 crc kubenswrapper[4914]: E1122 14:43:11.252578 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-547856594f-9j8pz: configmap "swift-ring-files" not found Nov 22 14:43:11 crc kubenswrapper[4914]: E1122 14:43:11.252636 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift podName:f7779691-44d2-40cd-8138-03953251ba7c nodeName:}" failed. No retries permitted until 2025-11-22 14:43:13.252618194 +0000 UTC m=+959.061187855 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift") pod "swift-proxy-547856594f-9j8pz" (UID: "f7779691-44d2-40cd-8138-03953251ba7c") : configmap "swift-ring-files" not found Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.353705 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-swiftconf\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.353779 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-dispersionconf\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.353800 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-scripts\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.353840 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-etc-swift\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.353856 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hflnp\" (UniqueName: \"kubernetes.io/projected/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-kube-api-access-hflnp\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.353895 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-ring-data-devices\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455071 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-dispersionconf\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455114 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-scripts\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455159 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-etc-swift\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455177 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hflnp\" (UniqueName: \"kubernetes.io/projected/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-kube-api-access-hflnp\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455230 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-ring-data-devices\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455288 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-swiftconf\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.455838 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-etc-swift\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.456235 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-ring-data-devices\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.456264 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-scripts\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.464462 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-swiftconf\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.464460 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-dispersionconf\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.476033 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hflnp\" (UniqueName: \"kubernetes.io/projected/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-kube-api-access-hflnp\") pod \"swift-ring-rebalance-zfxrw\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.540994 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.678921 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bggm4" event={"ID":"293f2650-2180-4730-9572-c5c59e370c97","Type":"ContainerStarted","Data":"c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1"} Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.706213 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-bggm4" podStartSLOduration=2.775419 podStartE2EDuration="4.706193835s" podCreationTimestamp="2025-11-22 14:43:07 +0000 UTC" firstStartedPulling="2025-11-22 14:43:08.685577389 +0000 UTC m=+954.494147050" lastFinishedPulling="2025-11-22 14:43:10.616352224 +0000 UTC m=+956.424921885" observedRunningTime="2025-11-22 14:43:11.701344016 +0000 UTC m=+957.509913697" watchObservedRunningTime="2025-11-22 14:43:11.706193835 +0000 UTC m=+957.514763496" Nov 22 14:43:11 crc kubenswrapper[4914]: I1122 14:43:11.809488 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-zfxrw"] Nov 22 14:43:11 crc kubenswrapper[4914]: W1122 14:43:11.816989 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9b69042_3ed9_4df0_86d9_70d4fd48bdc4.slice/crio-fda257f25ad3765efa761eed499a042fc9efd01a23a000489a2ec65863fc5011 WatchSource:0}: Error finding container fda257f25ad3765efa761eed499a042fc9efd01a23a000489a2ec65863fc5011: Status 404 returned error can't find the container with id fda257f25ad3765efa761eed499a042fc9efd01a23a000489a2ec65863fc5011 Nov 22 14:43:12 crc kubenswrapper[4914]: I1122 14:43:12.301389 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-bggm4"] Nov 22 14:43:12 crc kubenswrapper[4914]: I1122 14:43:12.689595 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" event={"ID":"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4","Type":"ContainerStarted","Data":"fda257f25ad3765efa761eed499a042fc9efd01a23a000489a2ec65863fc5011"} Nov 22 14:43:12 crc kubenswrapper[4914]: I1122 14:43:12.908304 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-5jsh9"] Nov 22 14:43:12 crc kubenswrapper[4914]: I1122 14:43:12.909475 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:12 crc kubenswrapper[4914]: I1122 14:43:12.917067 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-5jsh9"] Nov 22 14:43:12 crc kubenswrapper[4914]: I1122 14:43:12.980739 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxkt7\" (UniqueName: \"kubernetes.io/projected/2daccf25-7333-4068-9666-765c0e09e4f5-kube-api-access-fxkt7\") pod \"glance-operator-index-5jsh9\" (UID: \"2daccf25-7333-4068-9666-765c0e09e4f5\") " pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:13 crc kubenswrapper[4914]: I1122 14:43:13.082839 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxkt7\" (UniqueName: \"kubernetes.io/projected/2daccf25-7333-4068-9666-765c0e09e4f5-kube-api-access-fxkt7\") pod \"glance-operator-index-5jsh9\" (UID: \"2daccf25-7333-4068-9666-765c0e09e4f5\") " pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:13 crc kubenswrapper[4914]: I1122 14:43:13.109682 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxkt7\" (UniqueName: \"kubernetes.io/projected/2daccf25-7333-4068-9666-765c0e09e4f5-kube-api-access-fxkt7\") pod \"glance-operator-index-5jsh9\" (UID: \"2daccf25-7333-4068-9666-765c0e09e4f5\") " pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:13 crc kubenswrapper[4914]: I1122 14:43:13.271391 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:13 crc kubenswrapper[4914]: I1122 14:43:13.286566 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:13 crc kubenswrapper[4914]: E1122 14:43:13.286763 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:13 crc kubenswrapper[4914]: E1122 14:43:13.286794 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-547856594f-9j8pz: configmap "swift-ring-files" not found Nov 22 14:43:13 crc kubenswrapper[4914]: E1122 14:43:13.286855 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift podName:f7779691-44d2-40cd-8138-03953251ba7c nodeName:}" failed. No retries permitted until 2025-11-22 14:43:17.286836254 +0000 UTC m=+963.095405915 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift") pod "swift-proxy-547856594f-9j8pz" (UID: "f7779691-44d2-40cd-8138-03953251ba7c") : configmap "swift-ring-files" not found Nov 22 14:43:13 crc kubenswrapper[4914]: I1122 14:43:13.698898 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-bggm4" podUID="293f2650-2180-4730-9572-c5c59e370c97" containerName="registry-server" containerID="cri-o://c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1" gracePeriod=2 Nov 22 14:43:13 crc kubenswrapper[4914]: I1122 14:43:13.702380 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-5jsh9"] Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.038492 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.211624 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9st4v\" (UniqueName: \"kubernetes.io/projected/293f2650-2180-4730-9572-c5c59e370c97-kube-api-access-9st4v\") pod \"293f2650-2180-4730-9572-c5c59e370c97\" (UID: \"293f2650-2180-4730-9572-c5c59e370c97\") " Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.220734 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/293f2650-2180-4730-9572-c5c59e370c97-kube-api-access-9st4v" (OuterVolumeSpecName: "kube-api-access-9st4v") pod "293f2650-2180-4730-9572-c5c59e370c97" (UID: "293f2650-2180-4730-9572-c5c59e370c97"). InnerVolumeSpecName "kube-api-access-9st4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.314220 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9st4v\" (UniqueName: \"kubernetes.io/projected/293f2650-2180-4730-9572-c5c59e370c97-kube-api-access-9st4v\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.709349 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-5jsh9" event={"ID":"2daccf25-7333-4068-9666-765c0e09e4f5","Type":"ContainerStarted","Data":"815dd8c6f732d25eefac195e68f74087ad83df511b500d41cf01f955a50b663f"} Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.709391 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-5jsh9" event={"ID":"2daccf25-7333-4068-9666-765c0e09e4f5","Type":"ContainerStarted","Data":"399ac9ed585ff6ef1ae54be94bac6c3de1445072998d1c867123ca5be3c164d5"} Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.713276 4914 generic.go:334] "Generic (PLEG): container finished" podID="293f2650-2180-4730-9572-c5c59e370c97" containerID="c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1" exitCode=0 Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.713326 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-bggm4" Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.713324 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bggm4" event={"ID":"293f2650-2180-4730-9572-c5c59e370c97","Type":"ContainerDied","Data":"c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1"} Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.713449 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-bggm4" event={"ID":"293f2650-2180-4730-9572-c5c59e370c97","Type":"ContainerDied","Data":"3a32f2f38b508313d276eec1588bc5348380b018b45404bbcf4bfef86309f1b7"} Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.713470 4914 scope.go:117] "RemoveContainer" containerID="c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1" Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.729007 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-5jsh9" podStartSLOduration=2.662686446 podStartE2EDuration="2.728988754s" podCreationTimestamp="2025-11-22 14:43:12 +0000 UTC" firstStartedPulling="2025-11-22 14:43:13.711496579 +0000 UTC m=+959.520066240" lastFinishedPulling="2025-11-22 14:43:13.777798887 +0000 UTC m=+959.586368548" observedRunningTime="2025-11-22 14:43:14.721281365 +0000 UTC m=+960.529851026" watchObservedRunningTime="2025-11-22 14:43:14.728988754 +0000 UTC m=+960.537558415" Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.743907 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-bggm4"] Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.748037 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-bggm4"] Nov 22 14:43:14 crc kubenswrapper[4914]: I1122 14:43:14.877368 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="293f2650-2180-4730-9572-c5c59e370c97" path="/var/lib/kubelet/pods/293f2650-2180-4730-9572-c5c59e370c97/volumes" Nov 22 14:43:15 crc kubenswrapper[4914]: I1122 14:43:15.026468 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:15 crc kubenswrapper[4914]: E1122 14:43:15.026716 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:15 crc kubenswrapper[4914]: E1122 14:43:15.026749 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 22 14:43:15 crc kubenswrapper[4914]: E1122 14:43:15.026826 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift podName:cfa8c85b-6b20-4bde-a701-06b713f81c06 nodeName:}" failed. No retries permitted until 2025-11-22 14:43:23.026799212 +0000 UTC m=+968.835368883 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift") pod "swift-storage-0" (UID: "cfa8c85b-6b20-4bde-a701-06b713f81c06") : configmap "swift-ring-files" not found Nov 22 14:43:17 crc kubenswrapper[4914]: I1122 14:43:17.361655 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:17 crc kubenswrapper[4914]: E1122 14:43:17.361820 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:17 crc kubenswrapper[4914]: E1122 14:43:17.362025 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-547856594f-9j8pz: configmap "swift-ring-files" not found Nov 22 14:43:17 crc kubenswrapper[4914]: E1122 14:43:17.362075 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift podName:f7779691-44d2-40cd-8138-03953251ba7c nodeName:}" failed. No retries permitted until 2025-11-22 14:43:25.36206082 +0000 UTC m=+971.170630481 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift") pod "swift-proxy-547856594f-9j8pz" (UID: "f7779691-44d2-40cd-8138-03953251ba7c") : configmap "swift-ring-files" not found Nov 22 14:43:17 crc kubenswrapper[4914]: I1122 14:43:17.727402 4914 scope.go:117] "RemoveContainer" containerID="c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1" Nov 22 14:43:17 crc kubenswrapper[4914]: E1122 14:43:17.728164 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1\": container with ID starting with c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1 not found: ID does not exist" containerID="c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1" Nov 22 14:43:17 crc kubenswrapper[4914]: I1122 14:43:17.728199 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1"} err="failed to get container status \"c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1\": rpc error: code = NotFound desc = could not find container \"c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1\": container with ID starting with c7b215cfccb68c7fa3cb080825261e5e6f7f3ed31cf03126664bb109e4fb1db1 not found: ID does not exist" Nov 22 14:43:19 crc kubenswrapper[4914]: I1122 14:43:19.753050 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" event={"ID":"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4","Type":"ContainerStarted","Data":"baf2861b52b84c02a8181e1cd530696e449a4bb6e841a41bb70be1404278229c"} Nov 22 14:43:19 crc kubenswrapper[4914]: I1122 14:43:19.769438 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" podStartSLOduration=1.7375944479999998 podStartE2EDuration="8.769423518s" podCreationTimestamp="2025-11-22 14:43:11 +0000 UTC" firstStartedPulling="2025-11-22 14:43:11.819218173 +0000 UTC m=+957.627787834" lastFinishedPulling="2025-11-22 14:43:18.851047243 +0000 UTC m=+964.659616904" observedRunningTime="2025-11-22 14:43:19.768771112 +0000 UTC m=+965.577340773" watchObservedRunningTime="2025-11-22 14:43:19.769423518 +0000 UTC m=+965.577993179" Nov 22 14:43:23 crc kubenswrapper[4914]: I1122 14:43:23.046302 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:23 crc kubenswrapper[4914]: E1122 14:43:23.047071 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:23 crc kubenswrapper[4914]: E1122 14:43:23.047090 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 22 14:43:23 crc kubenswrapper[4914]: E1122 14:43:23.047145 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift podName:cfa8c85b-6b20-4bde-a701-06b713f81c06 nodeName:}" failed. No retries permitted until 2025-11-22 14:43:39.047127019 +0000 UTC m=+984.855696700 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift") pod "swift-storage-0" (UID: "cfa8c85b-6b20-4bde-a701-06b713f81c06") : configmap "swift-ring-files" not found Nov 22 14:43:23 crc kubenswrapper[4914]: I1122 14:43:23.272115 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:23 crc kubenswrapper[4914]: I1122 14:43:23.272181 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:23 crc kubenswrapper[4914]: I1122 14:43:23.296081 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:23 crc kubenswrapper[4914]: I1122 14:43:23.816642 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-5jsh9" Nov 22 14:43:25 crc kubenswrapper[4914]: I1122 14:43:25.398686 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:25 crc kubenswrapper[4914]: E1122 14:43:25.399420 4914 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 22 14:43:25 crc kubenswrapper[4914]: E1122 14:43:25.399450 4914 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-547856594f-9j8pz: configmap "swift-ring-files" not found Nov 22 14:43:25 crc kubenswrapper[4914]: E1122 14:43:25.399521 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift podName:f7779691-44d2-40cd-8138-03953251ba7c nodeName:}" failed. No retries permitted until 2025-11-22 14:43:41.399497634 +0000 UTC m=+987.208067305 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift") pod "swift-proxy-547856594f-9j8pz" (UID: "f7779691-44d2-40cd-8138-03953251ba7c") : configmap "swift-ring-files" not found Nov 22 14:43:25 crc kubenswrapper[4914]: I1122 14:43:25.797583 4914 generic.go:334] "Generic (PLEG): container finished" podID="e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" containerID="baf2861b52b84c02a8181e1cd530696e449a4bb6e841a41bb70be1404278229c" exitCode=0 Nov 22 14:43:25 crc kubenswrapper[4914]: I1122 14:43:25.797660 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" event={"ID":"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4","Type":"ContainerDied","Data":"baf2861b52b84c02a8181e1cd530696e449a4bb6e841a41bb70be1404278229c"} Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.360283 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz"] Nov 22 14:43:26 crc kubenswrapper[4914]: E1122 14:43:26.360583 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293f2650-2180-4730-9572-c5c59e370c97" containerName="registry-server" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.360596 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="293f2650-2180-4730-9572-c5c59e370c97" containerName="registry-server" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.360756 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="293f2650-2180-4730-9572-c5c59e370c97" containerName="registry-server" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.361864 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.364624 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l5bmb" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.387974 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz"] Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.518465 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r25lw\" (UniqueName: \"kubernetes.io/projected/4d27e00b-2931-49a6-9921-27c550ebe6e9-kube-api-access-r25lw\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.518637 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-bundle\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.518693 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-util\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.620079 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-bundle\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.620173 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-util\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.620255 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r25lw\" (UniqueName: \"kubernetes.io/projected/4d27e00b-2931-49a6-9921-27c550ebe6e9-kube-api-access-r25lw\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.620852 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-bundle\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.620891 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-util\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.644784 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r25lw\" (UniqueName: \"kubernetes.io/projected/4d27e00b-2931-49a6-9921-27c550ebe6e9-kube-api-access-r25lw\") pod \"9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:26 crc kubenswrapper[4914]: I1122 14:43:26.681074 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.072102 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.122193 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz"] Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.231256 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-swiftconf\") pod \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.231744 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-dispersionconf\") pod \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.231863 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-ring-data-devices\") pod \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.231913 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-scripts\") pod \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.231997 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-etc-swift\") pod \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.232041 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hflnp\" (UniqueName: \"kubernetes.io/projected/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-kube-api-access-hflnp\") pod \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\" (UID: \"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4\") " Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.232362 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" (UID: "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.232690 4914 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.233356 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" (UID: "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.237497 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-kube-api-access-hflnp" (OuterVolumeSpecName: "kube-api-access-hflnp") pod "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" (UID: "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4"). InnerVolumeSpecName "kube-api-access-hflnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.250494 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-scripts" (OuterVolumeSpecName: "scripts") pod "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" (UID: "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.252845 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" (UID: "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.254431 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" (UID: "e9b69042-3ed9-4df0-86d9-70d4fd48bdc4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.334445 4914 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.334479 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hflnp\" (UniqueName: \"kubernetes.io/projected/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-kube-api-access-hflnp\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.334489 4914 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.334498 4914 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.334507 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9b69042-3ed9-4df0-86d9-70d4fd48bdc4-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.820623 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" event={"ID":"e9b69042-3ed9-4df0-86d9-70d4fd48bdc4","Type":"ContainerDied","Data":"fda257f25ad3765efa761eed499a042fc9efd01a23a000489a2ec65863fc5011"} Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.820666 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fda257f25ad3765efa761eed499a042fc9efd01a23a000489a2ec65863fc5011" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.820686 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zfxrw" Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.825041 4914 generic.go:334] "Generic (PLEG): container finished" podID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerID="b2c830d9ab3116bfdfb1a6d94f9439476772c41f56cd22d4891a814bb0523768" exitCode=0 Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.825085 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" event={"ID":"4d27e00b-2931-49a6-9921-27c550ebe6e9","Type":"ContainerDied","Data":"b2c830d9ab3116bfdfb1a6d94f9439476772c41f56cd22d4891a814bb0523768"} Nov 22 14:43:27 crc kubenswrapper[4914]: I1122 14:43:27.825113 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" event={"ID":"4d27e00b-2931-49a6-9921-27c550ebe6e9","Type":"ContainerStarted","Data":"b02bdda7bc82d932c2880be953bc3ca0325ab0ba88499247e8dbc4f4df5e7e0e"} Nov 22 14:43:28 crc kubenswrapper[4914]: I1122 14:43:28.845525 4914 generic.go:334] "Generic (PLEG): container finished" podID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerID="d3fc43826a6e12bc81d6928ff2f48e179cb54351c70ae2f640f0fe968e639c72" exitCode=0 Nov 22 14:43:28 crc kubenswrapper[4914]: I1122 14:43:28.845632 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" event={"ID":"4d27e00b-2931-49a6-9921-27c550ebe6e9","Type":"ContainerDied","Data":"d3fc43826a6e12bc81d6928ff2f48e179cb54351c70ae2f640f0fe968e639c72"} Nov 22 14:43:29 crc kubenswrapper[4914]: I1122 14:43:29.862112 4914 generic.go:334] "Generic (PLEG): container finished" podID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerID="4a5264ca8915deeae6f528d925b1a3a51c0217e26528b1b7d4d2e8e33cbb783f" exitCode=0 Nov 22 14:43:29 crc kubenswrapper[4914]: I1122 14:43:29.862192 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" event={"ID":"4d27e00b-2931-49a6-9921-27c550ebe6e9","Type":"ContainerDied","Data":"4a5264ca8915deeae6f528d925b1a3a51c0217e26528b1b7d4d2e8e33cbb783f"} Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.186055 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.307910 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-util\") pod \"4d27e00b-2931-49a6-9921-27c550ebe6e9\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.308003 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r25lw\" (UniqueName: \"kubernetes.io/projected/4d27e00b-2931-49a6-9921-27c550ebe6e9-kube-api-access-r25lw\") pod \"4d27e00b-2931-49a6-9921-27c550ebe6e9\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.308112 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-bundle\") pod \"4d27e00b-2931-49a6-9921-27c550ebe6e9\" (UID: \"4d27e00b-2931-49a6-9921-27c550ebe6e9\") " Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.308827 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-bundle" (OuterVolumeSpecName: "bundle") pod "4d27e00b-2931-49a6-9921-27c550ebe6e9" (UID: "4d27e00b-2931-49a6-9921-27c550ebe6e9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.314062 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d27e00b-2931-49a6-9921-27c550ebe6e9-kube-api-access-r25lw" (OuterVolumeSpecName: "kube-api-access-r25lw") pod "4d27e00b-2931-49a6-9921-27c550ebe6e9" (UID: "4d27e00b-2931-49a6-9921-27c550ebe6e9"). InnerVolumeSpecName "kube-api-access-r25lw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.325999 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-util" (OuterVolumeSpecName: "util") pod "4d27e00b-2931-49a6-9921-27c550ebe6e9" (UID: "4d27e00b-2931-49a6-9921-27c550ebe6e9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.410651 4914 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.410713 4914 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d27e00b-2931-49a6-9921-27c550ebe6e9-util\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.410732 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r25lw\" (UniqueName: \"kubernetes.io/projected/4d27e00b-2931-49a6-9921-27c550ebe6e9-kube-api-access-r25lw\") on node \"crc\" DevicePath \"\"" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.892899 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" event={"ID":"4d27e00b-2931-49a6-9921-27c550ebe6e9","Type":"ContainerDied","Data":"b02bdda7bc82d932c2880be953bc3ca0325ab0ba88499247e8dbc4f4df5e7e0e"} Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.893307 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b02bdda7bc82d932c2880be953bc3ca0325ab0ba88499247e8dbc4f4df5e7e0e" Nov 22 14:43:31 crc kubenswrapper[4914]: I1122 14:43:31.892978 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz" Nov 22 14:43:39 crc kubenswrapper[4914]: I1122 14:43:39.136943 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:39 crc kubenswrapper[4914]: I1122 14:43:39.145211 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cfa8c85b-6b20-4bde-a701-06b713f81c06-etc-swift\") pod \"swift-storage-0\" (UID: \"cfa8c85b-6b20-4bde-a701-06b713f81c06\") " pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:39 crc kubenswrapper[4914]: I1122 14:43:39.290732 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 22 14:43:39 crc kubenswrapper[4914]: I1122 14:43:39.721070 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 22 14:43:39 crc kubenswrapper[4914]: I1122 14:43:39.953403 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"486e74f73318456833b23c2d67290ed72a1a126a78190c189ee260f28e6af430"} Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.474139 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.482637 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f7779691-44d2-40cd-8138-03953251ba7c-etc-swift\") pod \"swift-proxy-547856594f-9j8pz\" (UID: \"f7779691-44d2-40cd-8138-03953251ba7c\") " pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.682833 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.973444 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"6c9e48905d88a74ba77f5ecc583bc215a7b57851ca2eff2a67128752b1391393"} Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.973868 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"3e82ae1a4ec566a6c0eba942aa4f4a661ade65e834c1f6f8e9e7707a59f84466"} Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.973904 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"1debe703ec5d8c2fa86489c0eaf9b06952be0bea54059569f24fb3fdfe5ab784"} Nov 22 14:43:41 crc kubenswrapper[4914]: I1122 14:43:41.973920 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"89ec70a47acc0d937d3519c0bb6d348b0187e17d2187bf9e1eb6076c053aeb85"} Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.072049 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-547856594f-9j8pz"] Nov 22 14:43:42 crc kubenswrapper[4914]: W1122 14:43:42.076212 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7779691_44d2_40cd_8138_03953251ba7c.slice/crio-8b96b4de41979e89d68a8bb197b4493d9760a237cef661c9f0f4a6c7ea35d75b WatchSource:0}: Error finding container 8b96b4de41979e89d68a8bb197b4493d9760a237cef661c9f0f4a6c7ea35d75b: Status 404 returned error can't find the container with id 8b96b4de41979e89d68a8bb197b4493d9760a237cef661c9f0f4a6c7ea35d75b Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.975072 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq"] Nov 22 14:43:42 crc kubenswrapper[4914]: E1122 14:43:42.976006 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="util" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.976023 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="util" Nov 22 14:43:42 crc kubenswrapper[4914]: E1122 14:43:42.976038 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" containerName="swift-ring-rebalance" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.976046 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" containerName="swift-ring-rebalance" Nov 22 14:43:42 crc kubenswrapper[4914]: E1122 14:43:42.976061 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="pull" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.976069 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="pull" Nov 22 14:43:42 crc kubenswrapper[4914]: E1122 14:43:42.976083 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="extract" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.976090 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="extract" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.976240 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9b69042-3ed9-4df0-86d9-70d4fd48bdc4" containerName="swift-ring-rebalance" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.976286 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d27e00b-2931-49a6-9921-27c550ebe6e9" containerName="extract" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.977187 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.978911 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-29bnx" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.979638 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.981376 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" event={"ID":"f7779691-44d2-40cd-8138-03953251ba7c","Type":"ContainerStarted","Data":"21454188bed3a79212b0d79e631e6ca079cb89345a6dc9be79ff3865736d2f41"} Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.981413 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" event={"ID":"f7779691-44d2-40cd-8138-03953251ba7c","Type":"ContainerStarted","Data":"8b96b4de41979e89d68a8bb197b4493d9760a237cef661c9f0f4a6c7ea35d75b"} Nov 22 14:43:42 crc kubenswrapper[4914]: I1122 14:43:42.988003 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq"] Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.095619 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84f620d7-7666-4076-a141-b856b3548e6b-apiservice-cert\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.095692 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84f620d7-7666-4076-a141-b856b3548e6b-webhook-cert\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.095790 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8cjx\" (UniqueName: \"kubernetes.io/projected/84f620d7-7666-4076-a141-b856b3548e6b-kube-api-access-h8cjx\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.197036 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8cjx\" (UniqueName: \"kubernetes.io/projected/84f620d7-7666-4076-a141-b856b3548e6b-kube-api-access-h8cjx\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.197118 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84f620d7-7666-4076-a141-b856b3548e6b-apiservice-cert\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.197164 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84f620d7-7666-4076-a141-b856b3548e6b-webhook-cert\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.205201 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84f620d7-7666-4076-a141-b856b3548e6b-webhook-cert\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.205289 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84f620d7-7666-4076-a141-b856b3548e6b-apiservice-cert\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.334727 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8cjx\" (UniqueName: \"kubernetes.io/projected/84f620d7-7666-4076-a141-b856b3548e6b-kube-api-access-h8cjx\") pod \"glance-operator-controller-manager-bdf8b57db-6vrpq\" (UID: \"84f620d7-7666-4076-a141-b856b3548e6b\") " pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.349232 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:43 crc kubenswrapper[4914]: I1122 14:43:43.755248 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq"] Nov 22 14:43:43 crc kubenswrapper[4914]: W1122 14:43:43.756867 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84f620d7_7666_4076_a141_b856b3548e6b.slice/crio-b3c77e3efeb8fd3f4f48396eab721e34e6eaa7a788f4bd5e5ef458514339fb4f WatchSource:0}: Error finding container b3c77e3efeb8fd3f4f48396eab721e34e6eaa7a788f4bd5e5ef458514339fb4f: Status 404 returned error can't find the container with id b3c77e3efeb8fd3f4f48396eab721e34e6eaa7a788f4bd5e5ef458514339fb4f Nov 22 14:43:44 crc kubenswrapper[4914]: I1122 14:43:44.001556 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" event={"ID":"f7779691-44d2-40cd-8138-03953251ba7c","Type":"ContainerStarted","Data":"5a6c47fe5dc000ddf15b9c81f3570862c4b88b43f8c8c68251e83fe664eca709"} Nov 22 14:43:44 crc kubenswrapper[4914]: I1122 14:43:44.001613 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:44 crc kubenswrapper[4914]: I1122 14:43:44.001625 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:44 crc kubenswrapper[4914]: I1122 14:43:44.005141 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" event={"ID":"84f620d7-7666-4076-a141-b856b3548e6b","Type":"ContainerStarted","Data":"b3c77e3efeb8fd3f4f48396eab721e34e6eaa7a788f4bd5e5ef458514339fb4f"} Nov 22 14:43:44 crc kubenswrapper[4914]: I1122 14:43:44.024151 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" podStartSLOduration=35.024128442 podStartE2EDuration="35.024128442s" podCreationTimestamp="2025-11-22 14:43:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:43:44.017779217 +0000 UTC m=+989.826348898" watchObservedRunningTime="2025-11-22 14:43:44.024128442 +0000 UTC m=+989.832698103" Nov 22 14:43:45 crc kubenswrapper[4914]: I1122 14:43:45.025395 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"83c3145b5557ed1895ea65bed6da3577d7de32b4ed3e6f51825e311da04d4a4c"} Nov 22 14:43:45 crc kubenswrapper[4914]: I1122 14:43:45.025689 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"9a8ac930963c9088332328a43f009c6b9e31e477666b381b2ed649373e0da684"} Nov 22 14:43:46 crc kubenswrapper[4914]: I1122 14:43:46.041036 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"2fcb62b99149420338da7ded2626fa87bee149abae1ea957f407d1e5f516de8b"} Nov 22 14:43:46 crc kubenswrapper[4914]: I1122 14:43:46.041421 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"8de830591608aee2a23706501103a8f74380180e8693258e137b0ee5436eab2f"} Nov 22 14:43:46 crc kubenswrapper[4914]: I1122 14:43:46.043360 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" event={"ID":"84f620d7-7666-4076-a141-b856b3548e6b","Type":"ContainerStarted","Data":"821b711f841acfa0eaeb526e272471aed7bddf6069f808a2fec4381eda61f17a"} Nov 22 14:43:47 crc kubenswrapper[4914]: I1122 14:43:47.054969 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" event={"ID":"84f620d7-7666-4076-a141-b856b3548e6b","Type":"ContainerStarted","Data":"547ecbb42397d7d6f6ebc0c5614b4033c8e403fe0f02f53fc00be3a33a2dbf55"} Nov 22 14:43:47 crc kubenswrapper[4914]: I1122 14:43:47.055339 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:47 crc kubenswrapper[4914]: I1122 14:43:47.075433 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" podStartSLOduration=2.499776241 podStartE2EDuration="5.075413966s" podCreationTimestamp="2025-11-22 14:43:42 +0000 UTC" firstStartedPulling="2025-11-22 14:43:43.759145024 +0000 UTC m=+989.567714685" lastFinishedPulling="2025-11-22 14:43:46.334782729 +0000 UTC m=+992.143352410" observedRunningTime="2025-11-22 14:43:47.07192826 +0000 UTC m=+992.880497961" watchObservedRunningTime="2025-11-22 14:43:47.075413966 +0000 UTC m=+992.883983627" Nov 22 14:43:48 crc kubenswrapper[4914]: I1122 14:43:48.090302 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"e5f08db863fa76ef2b8fb3818c68f55646cfac4891675a93b24f4346950a829b"} Nov 22 14:43:48 crc kubenswrapper[4914]: I1122 14:43:48.090656 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"78b1304c7a008ec6fed7c65ed8f4dc44e7301823f16ecc479ed101657fb04e6d"} Nov 22 14:43:48 crc kubenswrapper[4914]: I1122 14:43:48.090672 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"7e92d3cf03fd61b160dc2dbc1ea03cf856d043579a39659a86da928404505719"} Nov 22 14:43:49 crc kubenswrapper[4914]: I1122 14:43:49.104344 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"712ee2f9e0a60a30a1a21de44e4de352a6cfc3cf545557cbb543473364de05ee"} Nov 22 14:43:49 crc kubenswrapper[4914]: I1122 14:43:49.104417 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"9931405392a21a2bb302fcaf4c61a96f779ad8efe82579eb242fac7cba3afe85"} Nov 22 14:43:49 crc kubenswrapper[4914]: I1122 14:43:49.104438 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"0d238642145f0ab8b93be5596ae010035dfcd575950e9a18e854e2a8f17ee267"} Nov 22 14:43:49 crc kubenswrapper[4914]: I1122 14:43:49.104460 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"cfa8c85b-6b20-4bde-a701-06b713f81c06","Type":"ContainerStarted","Data":"7fb09c67f4c417cd99e1780893d6e414542aa108027bf8fec0dfb9338dcf05fb"} Nov 22 14:43:49 crc kubenswrapper[4914]: I1122 14:43:49.156985 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=35.831831293 podStartE2EDuration="43.15695243s" podCreationTimestamp="2025-11-22 14:43:06 +0000 UTC" firstStartedPulling="2025-11-22 14:43:39.731001418 +0000 UTC m=+985.539571079" lastFinishedPulling="2025-11-22 14:43:47.056122555 +0000 UTC m=+992.864692216" observedRunningTime="2025-11-22 14:43:49.149480438 +0000 UTC m=+994.958050149" watchObservedRunningTime="2025-11-22 14:43:49.15695243 +0000 UTC m=+994.965522131" Nov 22 14:43:51 crc kubenswrapper[4914]: I1122 14:43:51.686224 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:51 crc kubenswrapper[4914]: I1122 14:43:51.690003 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-547856594f-9j8pz" Nov 22 14:43:53 crc kubenswrapper[4914]: I1122 14:43:53.353811 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-bdf8b57db-6vrpq" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.463497 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.465457 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.468156 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.468306 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.469701 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-h4nlp" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.470363 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.477348 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.628460 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02127756-1430-4482-a784-07a1a19f8048-openstack-config-secret\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.629004 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-scripts\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.629401 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-config\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.629967 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd8j6\" (UniqueName: \"kubernetes.io/projected/02127756-1430-4482-a784-07a1a19f8048-kube-api-access-fd8j6\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.733222 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-config\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.733586 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd8j6\" (UniqueName: \"kubernetes.io/projected/02127756-1430-4482-a784-07a1a19f8048-kube-api-access-fd8j6\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.733692 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02127756-1430-4482-a784-07a1a19f8048-openstack-config-secret\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.733723 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-scripts\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.734786 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-scripts\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.734900 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-config\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.741661 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02127756-1430-4482-a784-07a1a19f8048-openstack-config-secret\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.760416 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd8j6\" (UniqueName: \"kubernetes.io/projected/02127756-1430-4482-a784-07a1a19f8048-kube-api-access-fd8j6\") pod \"openstackclient\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:56 crc kubenswrapper[4914]: I1122 14:43:56.788522 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 22 14:43:57 crc kubenswrapper[4914]: I1122 14:43:57.238111 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:43:58 crc kubenswrapper[4914]: I1122 14:43:58.174108 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"02127756-1430-4482-a784-07a1a19f8048","Type":"ContainerStarted","Data":"f554761b697e97b53c8b410133a6009595a213bc5b857d4658d937b00caacc3a"} Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.134125 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-s9skv"] Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.135325 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.151056 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5rrm\" (UniqueName: \"kubernetes.io/projected/fbbbf055-8d4a-415d-8df0-fd21026f2b91-kube-api-access-z5rrm\") pod \"glance-db-create-s9skv\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.151150 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbbbf055-8d4a-415d-8df0-fd21026f2b91-operator-scripts\") pod \"glance-db-create-s9skv\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.152215 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-s9skv"] Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.165993 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-d2e1-account-create-update-6hndm"] Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.167089 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.168967 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.186918 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d2e1-account-create-update-6hndm"] Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.252737 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5rrm\" (UniqueName: \"kubernetes.io/projected/fbbbf055-8d4a-415d-8df0-fd21026f2b91-kube-api-access-z5rrm\") pod \"glance-db-create-s9skv\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.252842 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0d036e-a030-40f7-aee6-8312a6f82c21-operator-scripts\") pod \"glance-d2e1-account-create-update-6hndm\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.253082 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbbbf055-8d4a-415d-8df0-fd21026f2b91-operator-scripts\") pod \"glance-db-create-s9skv\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.253308 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flz45\" (UniqueName: \"kubernetes.io/projected/af0d036e-a030-40f7-aee6-8312a6f82c21-kube-api-access-flz45\") pod \"glance-d2e1-account-create-update-6hndm\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.253680 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbbbf055-8d4a-415d-8df0-fd21026f2b91-operator-scripts\") pod \"glance-db-create-s9skv\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.278089 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5rrm\" (UniqueName: \"kubernetes.io/projected/fbbbf055-8d4a-415d-8df0-fd21026f2b91-kube-api-access-z5rrm\") pod \"glance-db-create-s9skv\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.353981 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flz45\" (UniqueName: \"kubernetes.io/projected/af0d036e-a030-40f7-aee6-8312a6f82c21-kube-api-access-flz45\") pod \"glance-d2e1-account-create-update-6hndm\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.354047 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0d036e-a030-40f7-aee6-8312a6f82c21-operator-scripts\") pod \"glance-d2e1-account-create-update-6hndm\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.355520 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0d036e-a030-40f7-aee6-8312a6f82c21-operator-scripts\") pod \"glance-d2e1-account-create-update-6hndm\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.383024 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flz45\" (UniqueName: \"kubernetes.io/projected/af0d036e-a030-40f7-aee6-8312a6f82c21-kube-api-access-flz45\") pod \"glance-d2e1-account-create-update-6hndm\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.465004 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:01 crc kubenswrapper[4914]: I1122 14:44:01.483271 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:05 crc kubenswrapper[4914]: I1122 14:44:05.221944 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"02127756-1430-4482-a784-07a1a19f8048","Type":"ContainerStarted","Data":"177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0"} Nov 22 14:44:05 crc kubenswrapper[4914]: I1122 14:44:05.244696 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.5868310220000001 podStartE2EDuration="9.244677722s" podCreationTimestamp="2025-11-22 14:43:56 +0000 UTC" firstStartedPulling="2025-11-22 14:43:57.250932884 +0000 UTC m=+1003.059502555" lastFinishedPulling="2025-11-22 14:44:04.908779594 +0000 UTC m=+1010.717349255" observedRunningTime="2025-11-22 14:44:05.244548488 +0000 UTC m=+1011.053118179" watchObservedRunningTime="2025-11-22 14:44:05.244677722 +0000 UTC m=+1011.053247383" Nov 22 14:44:05 crc kubenswrapper[4914]: W1122 14:44:05.273973 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbbbf055_8d4a_415d_8df0_fd21026f2b91.slice/crio-1c8416be16ff3a6a4c96dfeab3a4e0c06e78a7a94933ad32f1f6e0bec4e7eef7 WatchSource:0}: Error finding container 1c8416be16ff3a6a4c96dfeab3a4e0c06e78a7a94933ad32f1f6e0bec4e7eef7: Status 404 returned error can't find the container with id 1c8416be16ff3a6a4c96dfeab3a4e0c06e78a7a94933ad32f1f6e0bec4e7eef7 Nov 22 14:44:05 crc kubenswrapper[4914]: I1122 14:44:05.283796 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-s9skv"] Nov 22 14:44:05 crc kubenswrapper[4914]: I1122 14:44:05.320182 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d2e1-account-create-update-6hndm"] Nov 22 14:44:06 crc kubenswrapper[4914]: I1122 14:44:06.231350 4914 generic.go:334] "Generic (PLEG): container finished" podID="fbbbf055-8d4a-415d-8df0-fd21026f2b91" containerID="86a95b11c221f8bf340a236501c9ef17e941e13345ace7f8d3f0f51eaa7fa7e9" exitCode=0 Nov 22 14:44:06 crc kubenswrapper[4914]: I1122 14:44:06.231469 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-s9skv" event={"ID":"fbbbf055-8d4a-415d-8df0-fd21026f2b91","Type":"ContainerDied","Data":"86a95b11c221f8bf340a236501c9ef17e941e13345ace7f8d3f0f51eaa7fa7e9"} Nov 22 14:44:06 crc kubenswrapper[4914]: I1122 14:44:06.231544 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-s9skv" event={"ID":"fbbbf055-8d4a-415d-8df0-fd21026f2b91","Type":"ContainerStarted","Data":"1c8416be16ff3a6a4c96dfeab3a4e0c06e78a7a94933ad32f1f6e0bec4e7eef7"} Nov 22 14:44:06 crc kubenswrapper[4914]: I1122 14:44:06.233074 4914 generic.go:334] "Generic (PLEG): container finished" podID="af0d036e-a030-40f7-aee6-8312a6f82c21" containerID="e44a87848313ee28111e3807e717a0cb1cb6e1eeea0a1187bf19023f121f9c7c" exitCode=0 Nov 22 14:44:06 crc kubenswrapper[4914]: I1122 14:44:06.233141 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" event={"ID":"af0d036e-a030-40f7-aee6-8312a6f82c21","Type":"ContainerDied","Data":"e44a87848313ee28111e3807e717a0cb1cb6e1eeea0a1187bf19023f121f9c7c"} Nov 22 14:44:06 crc kubenswrapper[4914]: I1122 14:44:06.233196 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" event={"ID":"af0d036e-a030-40f7-aee6-8312a6f82c21","Type":"ContainerStarted","Data":"82a9d1fd4c28be6613eb0060183bc8f9eb9385814077255ae61c314549703ac6"} Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.657423 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.662707 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.851407 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbbbf055-8d4a-415d-8df0-fd21026f2b91-operator-scripts\") pod \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.851723 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flz45\" (UniqueName: \"kubernetes.io/projected/af0d036e-a030-40f7-aee6-8312a6f82c21-kube-api-access-flz45\") pod \"af0d036e-a030-40f7-aee6-8312a6f82c21\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.851858 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5rrm\" (UniqueName: \"kubernetes.io/projected/fbbbf055-8d4a-415d-8df0-fd21026f2b91-kube-api-access-z5rrm\") pod \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\" (UID: \"fbbbf055-8d4a-415d-8df0-fd21026f2b91\") " Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.851998 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0d036e-a030-40f7-aee6-8312a6f82c21-operator-scripts\") pod \"af0d036e-a030-40f7-aee6-8312a6f82c21\" (UID: \"af0d036e-a030-40f7-aee6-8312a6f82c21\") " Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.852213 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbbbf055-8d4a-415d-8df0-fd21026f2b91-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fbbbf055-8d4a-415d-8df0-fd21026f2b91" (UID: "fbbbf055-8d4a-415d-8df0-fd21026f2b91"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.852410 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbbbf055-8d4a-415d-8df0-fd21026f2b91-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.853065 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0d036e-a030-40f7-aee6-8312a6f82c21-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af0d036e-a030-40f7-aee6-8312a6f82c21" (UID: "af0d036e-a030-40f7-aee6-8312a6f82c21"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.856543 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0d036e-a030-40f7-aee6-8312a6f82c21-kube-api-access-flz45" (OuterVolumeSpecName: "kube-api-access-flz45") pod "af0d036e-a030-40f7-aee6-8312a6f82c21" (UID: "af0d036e-a030-40f7-aee6-8312a6f82c21"). InnerVolumeSpecName "kube-api-access-flz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.859187 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbbbf055-8d4a-415d-8df0-fd21026f2b91-kube-api-access-z5rrm" (OuterVolumeSpecName: "kube-api-access-z5rrm") pod "fbbbf055-8d4a-415d-8df0-fd21026f2b91" (UID: "fbbbf055-8d4a-415d-8df0-fd21026f2b91"). InnerVolumeSpecName "kube-api-access-z5rrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.953523 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flz45\" (UniqueName: \"kubernetes.io/projected/af0d036e-a030-40f7-aee6-8312a6f82c21-kube-api-access-flz45\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.953617 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5rrm\" (UniqueName: \"kubernetes.io/projected/fbbbf055-8d4a-415d-8df0-fd21026f2b91-kube-api-access-z5rrm\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:07 crc kubenswrapper[4914]: I1122 14:44:07.953649 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af0d036e-a030-40f7-aee6-8312a6f82c21-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:08 crc kubenswrapper[4914]: I1122 14:44:08.249997 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" Nov 22 14:44:08 crc kubenswrapper[4914]: I1122 14:44:08.249997 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d2e1-account-create-update-6hndm" event={"ID":"af0d036e-a030-40f7-aee6-8312a6f82c21","Type":"ContainerDied","Data":"82a9d1fd4c28be6613eb0060183bc8f9eb9385814077255ae61c314549703ac6"} Nov 22 14:44:08 crc kubenswrapper[4914]: I1122 14:44:08.250123 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a9d1fd4c28be6613eb0060183bc8f9eb9385814077255ae61c314549703ac6" Nov 22 14:44:08 crc kubenswrapper[4914]: I1122 14:44:08.250895 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-s9skv" event={"ID":"fbbbf055-8d4a-415d-8df0-fd21026f2b91","Type":"ContainerDied","Data":"1c8416be16ff3a6a4c96dfeab3a4e0c06e78a7a94933ad32f1f6e0bec4e7eef7"} Nov 22 14:44:08 crc kubenswrapper[4914]: I1122 14:44:08.250918 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c8416be16ff3a6a4c96dfeab3a4e0c06e78a7a94933ad32f1f6e0bec4e7eef7" Nov 22 14:44:08 crc kubenswrapper[4914]: I1122 14:44:08.250951 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-s9skv" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.382386 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-cm6d7"] Nov 22 14:44:11 crc kubenswrapper[4914]: E1122 14:44:11.383469 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbbbf055-8d4a-415d-8df0-fd21026f2b91" containerName="mariadb-database-create" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.383491 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbbbf055-8d4a-415d-8df0-fd21026f2b91" containerName="mariadb-database-create" Nov 22 14:44:11 crc kubenswrapper[4914]: E1122 14:44:11.383522 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0d036e-a030-40f7-aee6-8312a6f82c21" containerName="mariadb-account-create-update" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.383536 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0d036e-a030-40f7-aee6-8312a6f82c21" containerName="mariadb-account-create-update" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.383794 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbbbf055-8d4a-415d-8df0-fd21026f2b91" containerName="mariadb-database-create" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.383830 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0d036e-a030-40f7-aee6-8312a6f82c21" containerName="mariadb-account-create-update" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.384698 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.387920 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.388036 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-wgm4l" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.395491 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cm6d7"] Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.405268 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-db-sync-config-data\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.405459 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44fw4\" (UniqueName: \"kubernetes.io/projected/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-kube-api-access-44fw4\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.405724 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-config-data\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.507658 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-db-sync-config-data\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.507717 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44fw4\" (UniqueName: \"kubernetes.io/projected/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-kube-api-access-44fw4\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.507774 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-config-data\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.524637 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44fw4\" (UniqueName: \"kubernetes.io/projected/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-kube-api-access-44fw4\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.528505 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-config-data\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.529618 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-db-sync-config-data\") pod \"glance-db-sync-cm6d7\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.708049 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:11 crc kubenswrapper[4914]: I1122 14:44:11.993374 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cm6d7"] Nov 22 14:44:12 crc kubenswrapper[4914]: I1122 14:44:12.284757 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cm6d7" event={"ID":"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd","Type":"ContainerStarted","Data":"b88a19ebfed9f1492eb724290c965de33c2afcdc21edc6009c6a7f4212fb6d96"} Nov 22 14:44:26 crc kubenswrapper[4914]: I1122 14:44:26.399792 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cm6d7" event={"ID":"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd","Type":"ContainerStarted","Data":"7dd998c7d54cddf21070c91b89f3f1bba5aa6ada3d742f23365ac28c55f663c1"} Nov 22 14:44:26 crc kubenswrapper[4914]: I1122 14:44:26.421765 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-cm6d7" podStartSLOduration=2.479485643 podStartE2EDuration="15.421743151s" podCreationTimestamp="2025-11-22 14:44:11 +0000 UTC" firstStartedPulling="2025-11-22 14:44:12.013513452 +0000 UTC m=+1017.822083113" lastFinishedPulling="2025-11-22 14:44:24.95577095 +0000 UTC m=+1030.764340621" observedRunningTime="2025-11-22 14:44:26.414319429 +0000 UTC m=+1032.222889090" watchObservedRunningTime="2025-11-22 14:44:26.421743151 +0000 UTC m=+1032.230312832" Nov 22 14:44:32 crc kubenswrapper[4914]: I1122 14:44:32.454930 4914 generic.go:334] "Generic (PLEG): container finished" podID="7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" containerID="7dd998c7d54cddf21070c91b89f3f1bba5aa6ada3d742f23365ac28c55f663c1" exitCode=0 Nov 22 14:44:32 crc kubenswrapper[4914]: I1122 14:44:32.455054 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cm6d7" event={"ID":"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd","Type":"ContainerDied","Data":"7dd998c7d54cddf21070c91b89f3f1bba5aa6ada3d742f23365ac28c55f663c1"} Nov 22 14:44:33 crc kubenswrapper[4914]: I1122 14:44:33.809561 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:33 crc kubenswrapper[4914]: I1122 14:44:33.973272 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44fw4\" (UniqueName: \"kubernetes.io/projected/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-kube-api-access-44fw4\") pod \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " Nov 22 14:44:33 crc kubenswrapper[4914]: I1122 14:44:33.973319 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-config-data\") pod \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " Nov 22 14:44:33 crc kubenswrapper[4914]: I1122 14:44:33.973344 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-db-sync-config-data\") pod \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\" (UID: \"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd\") " Nov 22 14:44:33 crc kubenswrapper[4914]: I1122 14:44:33.978637 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" (UID: "7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:44:33 crc kubenswrapper[4914]: I1122 14:44:33.987027 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-kube-api-access-44fw4" (OuterVolumeSpecName: "kube-api-access-44fw4") pod "7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" (UID: "7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd"). InnerVolumeSpecName "kube-api-access-44fw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.011911 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-config-data" (OuterVolumeSpecName: "config-data") pod "7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" (UID: "7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.075024 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44fw4\" (UniqueName: \"kubernetes.io/projected/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-kube-api-access-44fw4\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.075066 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.075083 4914 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.471069 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-cm6d7" event={"ID":"7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd","Type":"ContainerDied","Data":"b88a19ebfed9f1492eb724290c965de33c2afcdc21edc6009c6a7f4212fb6d96"} Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.471363 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b88a19ebfed9f1492eb724290c965de33c2afcdc21edc6009c6a7f4212fb6d96" Nov 22 14:44:34 crc kubenswrapper[4914]: I1122 14:44:34.471137 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-cm6d7" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.695160 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 22 14:44:35 crc kubenswrapper[4914]: E1122 14:44:35.695421 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" containerName="glance-db-sync" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.695432 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" containerName="glance-db-sync" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.695561 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" containerName="glance-db-sync" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.696205 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.697454 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-wgm4l" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.702320 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.711725 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.725615 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805436 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-config-data\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805492 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-logs\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805508 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-nvme\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805658 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805698 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-sys\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805720 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805760 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-scripts\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805782 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-run\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805808 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805837 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805854 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-dev\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805885 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-httpd-run\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805946 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcqpc\" (UniqueName: \"kubernetes.io/projected/462598ab-87c8-4175-bd52-2ebcdde65218-kube-api-access-tcqpc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.805976 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-lib-modules\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.877518 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.878975 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.887933 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907424 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-logs\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907469 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-nvme\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907495 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907625 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-nvme\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907637 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907740 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-sys\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907767 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907819 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-scripts\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907844 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-run\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907888 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907930 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907951 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-dev\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.907976 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-httpd-run\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908026 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcqpc\" (UniqueName: \"kubernetes.io/projected/462598ab-87c8-4175-bd52-2ebcdde65218-kube-api-access-tcqpc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908059 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-logs\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908087 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-lib-modules\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908113 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-lib-modules\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908148 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-sys\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908229 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-config-data\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908405 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.908842 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.910396 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-run\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.910753 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-dev\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.910993 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-httpd-run\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.911172 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.913039 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-config-data\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.917007 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-scripts\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.934772 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcqpc\" (UniqueName: \"kubernetes.io/projected/462598ab-87c8-4175-bd52-2ebcdde65218-kube-api-access-tcqpc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.935438 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:35 crc kubenswrapper[4914]: I1122 14:44:35.935646 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.009602 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.009998 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-scripts\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010021 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010060 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-run\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010108 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010127 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-dev\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010142 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-config-data\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010156 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-sys\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010169 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-logs\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010196 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkxjg\" (UniqueName: \"kubernetes.io/projected/2c166eff-4a21-4714-9e3e-668e9d71acb1-kube-api-access-jkxjg\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010217 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010240 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-httpd-run\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010261 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-lib-modules\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.010278 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.016298 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111587 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111660 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-dev\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111682 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-config-data\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111697 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-sys\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111777 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-logs\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111799 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkxjg\" (UniqueName: \"kubernetes.io/projected/2c166eff-4a21-4714-9e3e-668e9d71acb1-kube-api-access-jkxjg\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111706 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.111751 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-dev\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112000 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-sys\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112312 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-logs\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112358 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112385 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-httpd-run\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112405 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-lib-modules\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112423 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112466 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112491 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-scripts\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112511 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112535 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-run\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112591 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-run\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112619 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-lib-modules\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112655 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-httpd-run\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112799 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.112809 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.114091 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.114102 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.118072 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-scripts\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.118163 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-config-data\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.132948 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.142028 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkxjg\" (UniqueName: \"kubernetes.io/projected/2c166eff-4a21-4714-9e3e-668e9d71acb1-kube-api-access-jkxjg\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.156732 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.191625 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.376920 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.476812 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 22 14:44:36 crc kubenswrapper[4914]: I1122 14:44:36.484477 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2c166eff-4a21-4714-9e3e-668e9d71acb1","Type":"ContainerStarted","Data":"f5e737bdf335cffa01830deb0f5494760cc0d5511cadd1f76577540c78af24c6"} Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.494500 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"462598ab-87c8-4175-bd52-2ebcdde65218","Type":"ContainerStarted","Data":"ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879"} Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.495163 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"462598ab-87c8-4175-bd52-2ebcdde65218","Type":"ContainerStarted","Data":"f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f"} Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.495185 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"462598ab-87c8-4175-bd52-2ebcdde65218","Type":"ContainerStarted","Data":"19f91964d8eedc3067a5aef11312cfb40459ac224520748caab0989cda81b994"} Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.496501 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2c166eff-4a21-4714-9e3e-668e9d71acb1","Type":"ContainerStarted","Data":"8bc5cff776920fa71e44608901ce5beeef135e7d1c0dc780eec13f6167b35d9b"} Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.496522 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2c166eff-4a21-4714-9e3e-668e9d71acb1","Type":"ContainerStarted","Data":"8786ea1f0c4f6cf58a4bd17e8e5b8022d66b26d6196d9d693e89e434f07566df"} Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.526805 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.5267881660000002 podStartE2EDuration="2.526788166s" podCreationTimestamp="2025-11-22 14:44:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:44:37.522705827 +0000 UTC m=+1043.331275538" watchObservedRunningTime="2025-11-22 14:44:37.526788166 +0000 UTC m=+1043.335357817" Nov 22 14:44:37 crc kubenswrapper[4914]: I1122 14:44:37.557818 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.557787363 podStartE2EDuration="3.557787363s" podCreationTimestamp="2025-11-22 14:44:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:44:37.552379641 +0000 UTC m=+1043.360949302" watchObservedRunningTime="2025-11-22 14:44:37.557787363 +0000 UTC m=+1043.366357064" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.017906 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.018634 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.062769 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.084181 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.192618 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.192722 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.223133 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.241355 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.574066 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.574165 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.574195 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:46 crc kubenswrapper[4914]: I1122 14:44:46.574206 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.570821 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.588657 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.588699 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.589014 4914 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.642217 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.646952 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.720742 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:48 crc kubenswrapper[4914]: I1122 14:44:48.843025 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:50 crc kubenswrapper[4914]: I1122 14:44:50.616687 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-log" containerID="cri-o://8786ea1f0c4f6cf58a4bd17e8e5b8022d66b26d6196d9d693e89e434f07566df" gracePeriod=30 Nov 22 14:44:50 crc kubenswrapper[4914]: I1122 14:44:50.616758 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-httpd" containerID="cri-o://8bc5cff776920fa71e44608901ce5beeef135e7d1c0dc780eec13f6167b35d9b" gracePeriod=30 Nov 22 14:44:50 crc kubenswrapper[4914]: I1122 14:44:50.627518 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.102:9292/healthcheck\": EOF" Nov 22 14:44:51 crc kubenswrapper[4914]: I1122 14:44:51.623786 4914 generic.go:334] "Generic (PLEG): container finished" podID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerID="8786ea1f0c4f6cf58a4bd17e8e5b8022d66b26d6196d9d693e89e434f07566df" exitCode=143 Nov 22 14:44:51 crc kubenswrapper[4914]: I1122 14:44:51.624118 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2c166eff-4a21-4714-9e3e-668e9d71acb1","Type":"ContainerDied","Data":"8786ea1f0c4f6cf58a4bd17e8e5b8022d66b26d6196d9d693e89e434f07566df"} Nov 22 14:44:52 crc kubenswrapper[4914]: I1122 14:44:52.042274 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:44:52 crc kubenswrapper[4914]: I1122 14:44:52.042334 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:44:56 crc kubenswrapper[4914]: I1122 14:44:56.676135 4914 generic.go:334] "Generic (PLEG): container finished" podID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerID="8bc5cff776920fa71e44608901ce5beeef135e7d1c0dc780eec13f6167b35d9b" exitCode=0 Nov 22 14:44:56 crc kubenswrapper[4914]: I1122 14:44:56.676187 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2c166eff-4a21-4714-9e3e-668e9d71acb1","Type":"ContainerDied","Data":"8bc5cff776920fa71e44608901ce5beeef135e7d1c0dc780eec13f6167b35d9b"} Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.280088 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342249 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-lib-modules\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342302 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-iscsi\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342373 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342423 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-scripts\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342444 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-logs\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342484 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-sys\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342535 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkxjg\" (UniqueName: \"kubernetes.io/projected/2c166eff-4a21-4714-9e3e-668e9d71acb1-kube-api-access-jkxjg\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342563 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-nvme\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342611 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-var-locks-brick\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342629 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-run\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342653 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-dev\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342672 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342702 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-config-data\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.342721 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-httpd-run\") pod \"2c166eff-4a21-4714-9e3e-668e9d71acb1\" (UID: \"2c166eff-4a21-4714-9e3e-668e9d71acb1\") " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.343434 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.343478 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.343500 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.344112 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.344200 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-dev" (OuterVolumeSpecName: "dev") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.344276 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.344311 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-run" (OuterVolumeSpecName: "run") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.344310 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-sys" (OuterVolumeSpecName: "sys") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.345001 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-logs" (OuterVolumeSpecName: "logs") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.349492 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-scripts" (OuterVolumeSpecName: "scripts") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.350410 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.350595 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.351602 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c166eff-4a21-4714-9e3e-668e9d71acb1-kube-api-access-jkxjg" (OuterVolumeSpecName: "kube-api-access-jkxjg") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "kube-api-access-jkxjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.395740 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-config-data" (OuterVolumeSpecName: "config-data") pod "2c166eff-4a21-4714-9e3e-668e9d71acb1" (UID: "2c166eff-4a21-4714-9e3e-668e9d71acb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444201 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444243 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444255 4914 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444265 4914 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444276 4914 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444294 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444307 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c166eff-4a21-4714-9e3e-668e9d71acb1-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444317 4914 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c166eff-4a21-4714-9e3e-668e9d71acb1-logs\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444326 4914 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-sys\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444335 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkxjg\" (UniqueName: \"kubernetes.io/projected/2c166eff-4a21-4714-9e3e-668e9d71acb1-kube-api-access-jkxjg\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444346 4914 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444355 4914 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444365 4914 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.444377 4914 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c166eff-4a21-4714-9e3e-668e9d71acb1-dev\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.460144 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.461119 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.545684 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.545713 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.684977 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2c166eff-4a21-4714-9e3e-668e9d71acb1","Type":"ContainerDied","Data":"f5e737bdf335cffa01830deb0f5494760cc0d5511cadd1f76577540c78af24c6"} Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.685016 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.685040 4914 scope.go:117] "RemoveContainer" containerID="8bc5cff776920fa71e44608901ce5beeef135e7d1c0dc780eec13f6167b35d9b" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.717738 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.723676 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.727484 4914 scope.go:117] "RemoveContainer" containerID="8786ea1f0c4f6cf58a4bd17e8e5b8022d66b26d6196d9d693e89e434f07566df" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.751694 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:57 crc kubenswrapper[4914]: E1122 14:44:57.752055 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-log" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.752082 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-log" Nov 22 14:44:57 crc kubenswrapper[4914]: E1122 14:44:57.752127 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-httpd" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.752137 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-httpd" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.752305 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-log" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.752334 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" containerName="glance-httpd" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.756105 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.759150 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849530 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc69r\" (UniqueName: \"kubernetes.io/projected/efaa844e-59c0-480c-8d6a-8e5f0554c77b-kube-api-access-wc69r\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849588 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-httpd-run\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849628 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-config-data\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849654 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849796 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-lib-modules\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849894 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849921 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-dev\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.849991 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-logs\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.850087 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-run\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.850133 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.850286 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-nvme\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.850337 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.850422 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-sys\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.850456 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-scripts\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951536 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951581 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-dev\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951609 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-logs\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951636 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951655 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-run\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951704 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-dev\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951713 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-nvme\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951760 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-run\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951785 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-nvme\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951795 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951825 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-sys\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951827 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951840 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-scripts\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951862 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-sys\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951931 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc69r\" (UniqueName: \"kubernetes.io/projected/efaa844e-59c0-480c-8d6a-8e5f0554c77b-kube-api-access-wc69r\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951945 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.951977 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-httpd-run\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952001 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-config-data\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952026 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952043 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-logs\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952057 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-lib-modules\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952172 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-lib-modules\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952251 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952339 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-httpd-run\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.952582 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.955626 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-scripts\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.956643 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-config-data\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.968959 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc69r\" (UniqueName: \"kubernetes.io/projected/efaa844e-59c0-480c-8d6a-8e5f0554c77b-kube-api-access-wc69r\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.979686 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:57 crc kubenswrapper[4914]: I1122 14:44:57.979793 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:58 crc kubenswrapper[4914]: I1122 14:44:58.079587 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:44:58 crc kubenswrapper[4914]: I1122 14:44:58.499169 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:44:58 crc kubenswrapper[4914]: I1122 14:44:58.690374 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"efaa844e-59c0-480c-8d6a-8e5f0554c77b","Type":"ContainerStarted","Data":"743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729"} Nov 22 14:44:58 crc kubenswrapper[4914]: I1122 14:44:58.690743 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"efaa844e-59c0-480c-8d6a-8e5f0554c77b","Type":"ContainerStarted","Data":"01a4be8a183a6e64a4cb22d773c9d3c5b9fdc85fa4082029b5d718b194d8a699"} Nov 22 14:44:58 crc kubenswrapper[4914]: I1122 14:44:58.883804 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c166eff-4a21-4714-9e3e-668e9d71acb1" path="/var/lib/kubelet/pods/2c166eff-4a21-4714-9e3e-668e9d71acb1/volumes" Nov 22 14:44:59 crc kubenswrapper[4914]: I1122 14:44:59.703381 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"efaa844e-59c0-480c-8d6a-8e5f0554c77b","Type":"ContainerStarted","Data":"07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921"} Nov 22 14:44:59 crc kubenswrapper[4914]: I1122 14:44:59.754461 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.754434107 podStartE2EDuration="2.754434107s" podCreationTimestamp="2025-11-22 14:44:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:44:59.730936143 +0000 UTC m=+1065.539505874" watchObservedRunningTime="2025-11-22 14:44:59.754434107 +0000 UTC m=+1065.563003808" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.148298 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc"] Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.149587 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.151201 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.151821 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.182203 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc"] Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.188857 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84b3767a-c897-4fbc-92c8-7b251940a8e2-secret-volume\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.188929 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b3767a-c897-4fbc-92c8-7b251940a8e2-config-volume\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.189056 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9h6t\" (UniqueName: \"kubernetes.io/projected/84b3767a-c897-4fbc-92c8-7b251940a8e2-kube-api-access-p9h6t\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.290155 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9h6t\" (UniqueName: \"kubernetes.io/projected/84b3767a-c897-4fbc-92c8-7b251940a8e2-kube-api-access-p9h6t\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.290261 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84b3767a-c897-4fbc-92c8-7b251940a8e2-secret-volume\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.290293 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b3767a-c897-4fbc-92c8-7b251940a8e2-config-volume\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.291214 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b3767a-c897-4fbc-92c8-7b251940a8e2-config-volume\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.302998 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84b3767a-c897-4fbc-92c8-7b251940a8e2-secret-volume\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.309360 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9h6t\" (UniqueName: \"kubernetes.io/projected/84b3767a-c897-4fbc-92c8-7b251940a8e2-kube-api-access-p9h6t\") pod \"collect-profiles-29397045-j9bwc\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.466575 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:00 crc kubenswrapper[4914]: I1122 14:45:00.750295 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc"] Nov 22 14:45:01 crc kubenswrapper[4914]: I1122 14:45:01.721397 4914 generic.go:334] "Generic (PLEG): container finished" podID="84b3767a-c897-4fbc-92c8-7b251940a8e2" containerID="6af4924924eee4eebe064ebd9f3b4725c17d02f509c4a6eb0774e4309778b129" exitCode=0 Nov 22 14:45:01 crc kubenswrapper[4914]: I1122 14:45:01.721508 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" event={"ID":"84b3767a-c897-4fbc-92c8-7b251940a8e2","Type":"ContainerDied","Data":"6af4924924eee4eebe064ebd9f3b4725c17d02f509c4a6eb0774e4309778b129"} Nov 22 14:45:01 crc kubenswrapper[4914]: I1122 14:45:01.721761 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" event={"ID":"84b3767a-c897-4fbc-92c8-7b251940a8e2","Type":"ContainerStarted","Data":"1fbcc9375bf022c9c9b4aaef3e3e2bfedd77dfeb8f1666dcdd7339501488ba26"} Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.025461 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.133989 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b3767a-c897-4fbc-92c8-7b251940a8e2-config-volume\") pod \"84b3767a-c897-4fbc-92c8-7b251940a8e2\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.134061 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9h6t\" (UniqueName: \"kubernetes.io/projected/84b3767a-c897-4fbc-92c8-7b251940a8e2-kube-api-access-p9h6t\") pod \"84b3767a-c897-4fbc-92c8-7b251940a8e2\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.134172 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84b3767a-c897-4fbc-92c8-7b251940a8e2-secret-volume\") pod \"84b3767a-c897-4fbc-92c8-7b251940a8e2\" (UID: \"84b3767a-c897-4fbc-92c8-7b251940a8e2\") " Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.135096 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84b3767a-c897-4fbc-92c8-7b251940a8e2-config-volume" (OuterVolumeSpecName: "config-volume") pod "84b3767a-c897-4fbc-92c8-7b251940a8e2" (UID: "84b3767a-c897-4fbc-92c8-7b251940a8e2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.142253 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b3767a-c897-4fbc-92c8-7b251940a8e2-kube-api-access-p9h6t" (OuterVolumeSpecName: "kube-api-access-p9h6t") pod "84b3767a-c897-4fbc-92c8-7b251940a8e2" (UID: "84b3767a-c897-4fbc-92c8-7b251940a8e2"). InnerVolumeSpecName "kube-api-access-p9h6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.142423 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84b3767a-c897-4fbc-92c8-7b251940a8e2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "84b3767a-c897-4fbc-92c8-7b251940a8e2" (UID: "84b3767a-c897-4fbc-92c8-7b251940a8e2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.238283 4914 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84b3767a-c897-4fbc-92c8-7b251940a8e2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.238326 4914 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b3767a-c897-4fbc-92c8-7b251940a8e2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.238340 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9h6t\" (UniqueName: \"kubernetes.io/projected/84b3767a-c897-4fbc-92c8-7b251940a8e2-kube-api-access-p9h6t\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.739436 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" event={"ID":"84b3767a-c897-4fbc-92c8-7b251940a8e2","Type":"ContainerDied","Data":"1fbcc9375bf022c9c9b4aaef3e3e2bfedd77dfeb8f1666dcdd7339501488ba26"} Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.739509 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fbcc9375bf022c9c9b4aaef3e3e2bfedd77dfeb8f1666dcdd7339501488ba26" Nov 22 14:45:03 crc kubenswrapper[4914]: I1122 14:45:03.739590 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397045-j9bwc" Nov 22 14:45:08 crc kubenswrapper[4914]: I1122 14:45:08.081272 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:08 crc kubenswrapper[4914]: I1122 14:45:08.081793 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:08 crc kubenswrapper[4914]: I1122 14:45:08.120800 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:08 crc kubenswrapper[4914]: I1122 14:45:08.151701 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:08 crc kubenswrapper[4914]: I1122 14:45:08.795110 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:08 crc kubenswrapper[4914]: I1122 14:45:08.795167 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:10 crc kubenswrapper[4914]: I1122 14:45:10.744201 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:10 crc kubenswrapper[4914]: I1122 14:45:10.752565 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:13 crc kubenswrapper[4914]: I1122 14:45:13.789773 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-index-vcfbn" podUID="cd925b6d-ee1d-4af2-a157-492958a1d6ce" containerName="registry-server" probeResult="failure" output=< Nov 22 14:45:13 crc kubenswrapper[4914]: timeout: failed to connect service ":50051" within 1s Nov 22 14:45:13 crc kubenswrapper[4914]: > Nov 22 14:45:13 crc kubenswrapper[4914]: I1122 14:45:13.805106 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-index-vcfbn" podUID="cd925b6d-ee1d-4af2-a157-492958a1d6ce" containerName="registry-server" probeResult="failure" output=< Nov 22 14:45:13 crc kubenswrapper[4914]: timeout: failed to connect service ":50051" within 1s Nov 22 14:45:13 crc kubenswrapper[4914]: > Nov 22 14:45:22 crc kubenswrapper[4914]: I1122 14:45:22.042556 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:45:22 crc kubenswrapper[4914]: I1122 14:45:22.042935 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.071383 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cm6d7"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.083447 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-cm6d7"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.103197 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanced2e1-account-delete-kw6m4"] Nov 22 14:45:27 crc kubenswrapper[4914]: E1122 14:45:27.103782 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b3767a-c897-4fbc-92c8-7b251940a8e2" containerName="collect-profiles" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.103913 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b3767a-c897-4fbc-92c8-7b251940a8e2" containerName="collect-profiles" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.104176 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b3767a-c897-4fbc-92c8-7b251940a8e2" containerName="collect-profiles" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.104808 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.124767 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced2e1-account-delete-kw6m4"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.160045 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.160358 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-log" containerID="cri-o://f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f" gracePeriod=30 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.160795 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-httpd" containerID="cri-o://ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879" gracePeriod=30 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.166833 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.167160 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-log" containerID="cri-o://743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729" gracePeriod=30 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.167326 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-httpd" containerID="cri-o://07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921" gracePeriod=30 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.207760 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-operator-scripts\") pod \"glanced2e1-account-delete-kw6m4\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.208060 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq5pd\" (UniqueName: \"kubernetes.io/projected/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-kube-api-access-gq5pd\") pod \"glanced2e1-account-delete-kw6m4\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.272559 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.272918 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="02127756-1430-4482-a784-07a1a19f8048" containerName="openstackclient" containerID="cri-o://177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0" gracePeriod=30 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.309055 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-operator-scripts\") pod \"glanced2e1-account-delete-kw6m4\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.309197 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq5pd\" (UniqueName: \"kubernetes.io/projected/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-kube-api-access-gq5pd\") pod \"glanced2e1-account-delete-kw6m4\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.309999 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-operator-scripts\") pod \"glanced2e1-account-delete-kw6m4\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.331710 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq5pd\" (UniqueName: \"kubernetes.io/projected/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-kube-api-access-gq5pd\") pod \"glanced2e1-account-delete-kw6m4\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.454948 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.693644 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.714971 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-scripts\") pod \"02127756-1430-4482-a784-07a1a19f8048\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.715078 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02127756-1430-4482-a784-07a1a19f8048-openstack-config-secret\") pod \"02127756-1430-4482-a784-07a1a19f8048\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.715132 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-config\") pod \"02127756-1430-4482-a784-07a1a19f8048\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.715177 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd8j6\" (UniqueName: \"kubernetes.io/projected/02127756-1430-4482-a784-07a1a19f8048-kube-api-access-fd8j6\") pod \"02127756-1430-4482-a784-07a1a19f8048\" (UID: \"02127756-1430-4482-a784-07a1a19f8048\") " Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.715542 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "02127756-1430-4482-a784-07a1a19f8048" (UID: "02127756-1430-4482-a784-07a1a19f8048"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.715965 4914 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.722998 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02127756-1430-4482-a784-07a1a19f8048-kube-api-access-fd8j6" (OuterVolumeSpecName: "kube-api-access-fd8j6") pod "02127756-1430-4482-a784-07a1a19f8048" (UID: "02127756-1430-4482-a784-07a1a19f8048"). InnerVolumeSpecName "kube-api-access-fd8j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.733769 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "02127756-1430-4482-a784-07a1a19f8048" (UID: "02127756-1430-4482-a784-07a1a19f8048"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.740616 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02127756-1430-4482-a784-07a1a19f8048-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "02127756-1430-4482-a784-07a1a19f8048" (UID: "02127756-1430-4482-a784-07a1a19f8048"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.817784 4914 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02127756-1430-4482-a784-07a1a19f8048-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.817814 4914 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02127756-1430-4482-a784-07a1a19f8048-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.817823 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd8j6\" (UniqueName: \"kubernetes.io/projected/02127756-1430-4482-a784-07a1a19f8048-kube-api-access-fd8j6\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.886683 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced2e1-account-delete-kw6m4"] Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.958197 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" event={"ID":"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168","Type":"ContainerStarted","Data":"e9a31942dc27a5b5a15922424fc50be2ebd7cb4b98433b0ccd8452acc41ecd93"} Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.959731 4914 generic.go:334] "Generic (PLEG): container finished" podID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerID="743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729" exitCode=143 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.959811 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"efaa844e-59c0-480c-8d6a-8e5f0554c77b","Type":"ContainerDied","Data":"743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729"} Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.961259 4914 generic.go:334] "Generic (PLEG): container finished" podID="02127756-1430-4482-a784-07a1a19f8048" containerID="177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0" exitCode=143 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.961301 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"02127756-1430-4482-a784-07a1a19f8048","Type":"ContainerDied","Data":"177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0"} Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.961320 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.961337 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"02127756-1430-4482-a784-07a1a19f8048","Type":"ContainerDied","Data":"f554761b697e97b53c8b410133a6009595a213bc5b857d4658d937b00caacc3a"} Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.961360 4914 scope.go:117] "RemoveContainer" containerID="177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.962984 4914 generic.go:334] "Generic (PLEG): container finished" podID="462598ab-87c8-4175-bd52-2ebcdde65218" containerID="f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f" exitCode=143 Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.963018 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"462598ab-87c8-4175-bd52-2ebcdde65218","Type":"ContainerDied","Data":"f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f"} Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.989969 4914 scope.go:117] "RemoveContainer" containerID="177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0" Nov 22 14:45:27 crc kubenswrapper[4914]: E1122 14:45:27.994210 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0\": container with ID starting with 177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0 not found: ID does not exist" containerID="177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0" Nov 22 14:45:27 crc kubenswrapper[4914]: I1122 14:45:27.994287 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0"} err="failed to get container status \"177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0\": rpc error: code = NotFound desc = could not find container \"177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0\": container with ID starting with 177ec3ad19399b0665053a304098d8eeb11546f13803125ecafab11149b417f0 not found: ID does not exist" Nov 22 14:45:28 crc kubenswrapper[4914]: I1122 14:45:28.009590 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:45:28 crc kubenswrapper[4914]: I1122 14:45:28.014257 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:45:28 crc kubenswrapper[4914]: I1122 14:45:28.881426 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02127756-1430-4482-a784-07a1a19f8048" path="/var/lib/kubelet/pods/02127756-1430-4482-a784-07a1a19f8048/volumes" Nov 22 14:45:28 crc kubenswrapper[4914]: I1122 14:45:28.882731 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd" path="/var/lib/kubelet/pods/7bd4c2bd-f24b-4ac5-b089-a0d818a5e6dd/volumes" Nov 22 14:45:28 crc kubenswrapper[4914]: I1122 14:45:28.989272 4914 generic.go:334] "Generic (PLEG): container finished" podID="6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" containerID="11f5f0d72a1007462d2b0f5f66d7f9643eb66229ccf9d41656b1517cc77cd893" exitCode=0 Nov 22 14:45:28 crc kubenswrapper[4914]: I1122 14:45:28.989326 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" event={"ID":"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168","Type":"ContainerDied","Data":"11f5f0d72a1007462d2b0f5f66d7f9643eb66229ccf9d41656b1517cc77cd893"} Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.326721 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.103:9292/healthcheck\": read tcp 10.217.0.2:46084->10.217.0.103:9292: read: connection reset by peer" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.329947 4914 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.103:9292/healthcheck\": read tcp 10.217.0.2:46094->10.217.0.103:9292: read: connection reset by peer" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.334484 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.477665 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-operator-scripts\") pod \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.477760 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq5pd\" (UniqueName: \"kubernetes.io/projected/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-kube-api-access-gq5pd\") pod \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\" (UID: \"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.478300 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" (UID: "6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.488588 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-kube-api-access-gq5pd" (OuterVolumeSpecName: "kube-api-access-gq5pd") pod "6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" (UID: "6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168"). InnerVolumeSpecName "kube-api-access-gq5pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: E1122 14:45:30.548339 4914 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefaa844e_59c0_480c_8d6a_8e5f0554c77b.slice/crio-conmon-07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921.scope\": RecentStats: unable to find data in memory cache]" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.588033 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.588089 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq5pd\" (UniqueName: \"kubernetes.io/projected/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168-kube-api-access-gq5pd\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.668765 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.673621 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791238 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-iscsi\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791309 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-run\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791357 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-nvme\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791395 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-logs\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791442 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-var-locks-brick\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791529 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791555 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791583 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-nvme\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791628 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-config-data\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791659 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-scripts\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791687 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-sys\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791714 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-httpd-run\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791756 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc69r\" (UniqueName: \"kubernetes.io/projected/efaa844e-59c0-480c-8d6a-8e5f0554c77b-kube-api-access-wc69r\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791781 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-dev\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791810 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-sys\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791839 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-dev\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791897 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcqpc\" (UniqueName: \"kubernetes.io/projected/462598ab-87c8-4175-bd52-2ebcdde65218-kube-api-access-tcqpc\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791916 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-lib-modules\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791950 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-lib-modules\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.791978 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-httpd-run\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792004 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792025 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-var-locks-brick\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792060 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-logs\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792092 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-scripts\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792117 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\" (UID: \"efaa844e-59c0-480c-8d6a-8e5f0554c77b\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792156 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-iscsi\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792186 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-run\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792187 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-sys" (OuterVolumeSpecName: "sys") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792214 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-sys" (OuterVolumeSpecName: "sys") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792284 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-run" (OuterVolumeSpecName: "run") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792305 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792322 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792249 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792250 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792478 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792662 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792740 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-dev" (OuterVolumeSpecName: "dev") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792781 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792812 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792819 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-logs" (OuterVolumeSpecName: "logs") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792841 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-dev" (OuterVolumeSpecName: "dev") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792856 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.792979 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-logs" (OuterVolumeSpecName: "logs") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.793045 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.793092 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-run" (OuterVolumeSpecName: "run") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.793271 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-config-data\") pod \"462598ab-87c8-4175-bd52-2ebcdde65218\" (UID: \"462598ab-87c8-4175-bd52-2ebcdde65218\") " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794088 4914 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794113 4914 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-sys\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794124 4914 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794134 4914 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-dev\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794145 4914 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-sys\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794156 4914 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-dev\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794167 4914 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794177 4914 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794187 4914 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794198 4914 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794209 4914 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462598ab-87c8-4175-bd52-2ebcdde65218-logs\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794219 4914 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794231 4914 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794241 4914 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794251 4914 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/efaa844e-59c0-480c-8d6a-8e5f0554c77b-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794263 4914 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794273 4914 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efaa844e-59c0-480c-8d6a-8e5f0554c77b-logs\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.794283 4914 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/462598ab-87c8-4175-bd52-2ebcdde65218-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.795739 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efaa844e-59c0-480c-8d6a-8e5f0554c77b-kube-api-access-wc69r" (OuterVolumeSpecName: "kube-api-access-wc69r") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "kube-api-access-wc69r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.795867 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.795994 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.796044 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.796223 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.796729 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462598ab-87c8-4175-bd52-2ebcdde65218-kube-api-access-tcqpc" (OuterVolumeSpecName: "kube-api-access-tcqpc") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "kube-api-access-tcqpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.796757 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-scripts" (OuterVolumeSpecName: "scripts") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.797071 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-scripts" (OuterVolumeSpecName: "scripts") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.827042 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-config-data" (OuterVolumeSpecName: "config-data") pod "462598ab-87c8-4175-bd52-2ebcdde65218" (UID: "462598ab-87c8-4175-bd52-2ebcdde65218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.839999 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-config-data" (OuterVolumeSpecName: "config-data") pod "efaa844e-59c0-480c-8d6a-8e5f0554c77b" (UID: "efaa844e-59c0-480c-8d6a-8e5f0554c77b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899573 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899615 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899632 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899646 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899659 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc69r\" (UniqueName: \"kubernetes.io/projected/efaa844e-59c0-480c-8d6a-8e5f0554c77b-kube-api-access-wc69r\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899673 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcqpc\" (UniqueName: \"kubernetes.io/projected/462598ab-87c8-4175-bd52-2ebcdde65218-kube-api-access-tcqpc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899691 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899703 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efaa844e-59c0-480c-8d6a-8e5f0554c77b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899720 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.899732 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462598ab-87c8-4175-bd52-2ebcdde65218-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.919358 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.925273 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.933834 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 22 14:45:30 crc kubenswrapper[4914]: I1122 14:45:30.946526 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.000296 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.000322 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.000331 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.000339 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.007944 4914 generic.go:334] "Generic (PLEG): container finished" podID="462598ab-87c8-4175-bd52-2ebcdde65218" containerID="ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879" exitCode=0 Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.008011 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.008011 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"462598ab-87c8-4175-bd52-2ebcdde65218","Type":"ContainerDied","Data":"ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879"} Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.008111 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"462598ab-87c8-4175-bd52-2ebcdde65218","Type":"ContainerDied","Data":"19f91964d8eedc3067a5aef11312cfb40459ac224520748caab0989cda81b994"} Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.008128 4914 scope.go:117] "RemoveContainer" containerID="ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.009547 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" event={"ID":"6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168","Type":"ContainerDied","Data":"e9a31942dc27a5b5a15922424fc50be2ebd7cb4b98433b0ccd8452acc41ecd93"} Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.009569 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9a31942dc27a5b5a15922424fc50be2ebd7cb4b98433b0ccd8452acc41ecd93" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.009608 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced2e1-account-delete-kw6m4" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.012752 4914 generic.go:334] "Generic (PLEG): container finished" podID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerID="07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921" exitCode=0 Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.012824 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"efaa844e-59c0-480c-8d6a-8e5f0554c77b","Type":"ContainerDied","Data":"07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921"} Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.012909 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"efaa844e-59c0-480c-8d6a-8e5f0554c77b","Type":"ContainerDied","Data":"01a4be8a183a6e64a4cb22d773c9d3c5b9fdc85fa4082029b5d718b194d8a699"} Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.012936 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.036686 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.042215 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.046418 4914 scope.go:117] "RemoveContainer" containerID="f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.052054 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.056644 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.064334 4914 scope.go:117] "RemoveContainer" containerID="ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879" Nov 22 14:45:31 crc kubenswrapper[4914]: E1122 14:45:31.064903 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879\": container with ID starting with ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879 not found: ID does not exist" containerID="ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.064936 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879"} err="failed to get container status \"ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879\": rpc error: code = NotFound desc = could not find container \"ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879\": container with ID starting with ff08ceb19041048416b66fa854fe02cfc5a3d68fed11348979c21e120fb4b879 not found: ID does not exist" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.064960 4914 scope.go:117] "RemoveContainer" containerID="f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f" Nov 22 14:45:31 crc kubenswrapper[4914]: E1122 14:45:31.065368 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f\": container with ID starting with f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f not found: ID does not exist" containerID="f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.065405 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f"} err="failed to get container status \"f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f\": rpc error: code = NotFound desc = could not find container \"f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f\": container with ID starting with f0dd951901981796275629110e01ce5369104575b78415b0ecec0526828aa97f not found: ID does not exist" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.065433 4914 scope.go:117] "RemoveContainer" containerID="07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.087333 4914 scope.go:117] "RemoveContainer" containerID="743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.113357 4914 scope.go:117] "RemoveContainer" containerID="07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921" Nov 22 14:45:31 crc kubenswrapper[4914]: E1122 14:45:31.115060 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921\": container with ID starting with 07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921 not found: ID does not exist" containerID="07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.115128 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921"} err="failed to get container status \"07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921\": rpc error: code = NotFound desc = could not find container \"07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921\": container with ID starting with 07de97bd76c4b0214cf5bc686c50f9bee56af4050354ec4f209c4a9cfc8c1921 not found: ID does not exist" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.115172 4914 scope.go:117] "RemoveContainer" containerID="743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729" Nov 22 14:45:31 crc kubenswrapper[4914]: E1122 14:45:31.115553 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729\": container with ID starting with 743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729 not found: ID does not exist" containerID="743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729" Nov 22 14:45:31 crc kubenswrapper[4914]: I1122 14:45:31.115600 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729"} err="failed to get container status \"743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729\": rpc error: code = NotFound desc = could not find container \"743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729\": container with ID starting with 743f49199dae46bbed6566b08c3e637798e3cf0e8bdfa1008407f1944153c729 not found: ID does not exist" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.142256 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-s9skv"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.149616 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-s9skv"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.171187 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-d2e1-account-create-update-6hndm"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.177579 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanced2e1-account-delete-kw6m4"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.184848 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-d2e1-account-create-update-6hndm"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.189698 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanced2e1-account-delete-kw6m4"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.219985 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-zbdng"] Nov 22 14:45:32 crc kubenswrapper[4914]: E1122 14:45:32.220263 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-httpd" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220276 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-httpd" Nov 22 14:45:32 crc kubenswrapper[4914]: E1122 14:45:32.220292 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-log" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220297 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-log" Nov 22 14:45:32 crc kubenswrapper[4914]: E1122 14:45:32.220314 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-httpd" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220321 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-httpd" Nov 22 14:45:32 crc kubenswrapper[4914]: E1122 14:45:32.220334 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-log" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220339 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-log" Nov 22 14:45:32 crc kubenswrapper[4914]: E1122 14:45:32.220346 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" containerName="mariadb-account-delete" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220352 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" containerName="mariadb-account-delete" Nov 22 14:45:32 crc kubenswrapper[4914]: E1122 14:45:32.220360 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02127756-1430-4482-a784-07a1a19f8048" containerName="openstackclient" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220365 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="02127756-1430-4482-a784-07a1a19f8048" containerName="openstackclient" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220512 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-log" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220526 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" containerName="glance-httpd" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220537 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" containerName="mariadb-account-delete" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220547 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-httpd" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220557 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" containerName="glance-log" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.220566 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="02127756-1430-4482-a784-07a1a19f8048" containerName="openstackclient" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.221051 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.240656 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbdng"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.319829 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctj8d\" (UniqueName: \"kubernetes.io/projected/9a4c03fc-56f1-4e81-975b-7e3baf922436-kube-api-access-ctj8d\") pod \"glance-db-create-zbdng\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.320225 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4c03fc-56f1-4e81-975b-7e3baf922436-operator-scripts\") pod \"glance-db-create-zbdng\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.335339 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.336740 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.343291 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.349685 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk"] Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.421623 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417645f9-c3c1-49c8-bcc4-84258eeb062d-operator-scripts\") pod \"glance-a3b4-account-create-update-qg5gk\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.421664 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48xjz\" (UniqueName: \"kubernetes.io/projected/417645f9-c3c1-49c8-bcc4-84258eeb062d-kube-api-access-48xjz\") pod \"glance-a3b4-account-create-update-qg5gk\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.421720 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctj8d\" (UniqueName: \"kubernetes.io/projected/9a4c03fc-56f1-4e81-975b-7e3baf922436-kube-api-access-ctj8d\") pod \"glance-db-create-zbdng\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.421964 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4c03fc-56f1-4e81-975b-7e3baf922436-operator-scripts\") pod \"glance-db-create-zbdng\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.422708 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4c03fc-56f1-4e81-975b-7e3baf922436-operator-scripts\") pod \"glance-db-create-zbdng\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.444689 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctj8d\" (UniqueName: \"kubernetes.io/projected/9a4c03fc-56f1-4e81-975b-7e3baf922436-kube-api-access-ctj8d\") pod \"glance-db-create-zbdng\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.524190 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417645f9-c3c1-49c8-bcc4-84258eeb062d-operator-scripts\") pod \"glance-a3b4-account-create-update-qg5gk\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.524266 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48xjz\" (UniqueName: \"kubernetes.io/projected/417645f9-c3c1-49c8-bcc4-84258eeb062d-kube-api-access-48xjz\") pod \"glance-a3b4-account-create-update-qg5gk\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.525411 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417645f9-c3c1-49c8-bcc4-84258eeb062d-operator-scripts\") pod \"glance-a3b4-account-create-update-qg5gk\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.542716 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.543343 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48xjz\" (UniqueName: \"kubernetes.io/projected/417645f9-c3c1-49c8-bcc4-84258eeb062d-kube-api-access-48xjz\") pod \"glance-a3b4-account-create-update-qg5gk\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:32 crc kubenswrapper[4914]: I1122 14:45:32.653131 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:32.878116 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462598ab-87c8-4175-bd52-2ebcdde65218" path="/var/lib/kubelet/pods/462598ab-87c8-4175-bd52-2ebcdde65218/volumes" Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:32.878979 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168" path="/var/lib/kubelet/pods/6b1bf461-4d69-4d27-8bfc-d0fa6f8ed168/volumes" Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:32.879746 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0d036e-a030-40f7-aee6-8312a6f82c21" path="/var/lib/kubelet/pods/af0d036e-a030-40f7-aee6-8312a6f82c21/volumes" Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:32.881120 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efaa844e-59c0-480c-8d6a-8e5f0554c77b" path="/var/lib/kubelet/pods/efaa844e-59c0-480c-8d6a-8e5f0554c77b/volumes" Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:32.881854 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbbbf055-8d4a-415d-8df0-fd21026f2b91" path="/var/lib/kubelet/pods/fbbbf055-8d4a-415d-8df0-fd21026f2b91/volumes" Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:32.994698 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbdng"] Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:33.032910 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zbdng" event={"ID":"9a4c03fc-56f1-4e81-975b-7e3baf922436","Type":"ContainerStarted","Data":"fa51523087edb91e734665c820215d3f8ed2fd349a17a5fbbaca70a3b72dc284"} Nov 22 14:45:33 crc kubenswrapper[4914]: W1122 14:45:33.779797 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod417645f9_c3c1_49c8_bcc4_84258eeb062d.slice/crio-56e9aa71fb08ca30002ab9f25d655c04631778ad2237186c7d1a1690acd48fff WatchSource:0}: Error finding container 56e9aa71fb08ca30002ab9f25d655c04631778ad2237186c7d1a1690acd48fff: Status 404 returned error can't find the container with id 56e9aa71fb08ca30002ab9f25d655c04631778ad2237186c7d1a1690acd48fff Nov 22 14:45:33 crc kubenswrapper[4914]: I1122 14:45:33.785241 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk"] Nov 22 14:45:34 crc kubenswrapper[4914]: I1122 14:45:34.042021 4914 generic.go:334] "Generic (PLEG): container finished" podID="9a4c03fc-56f1-4e81-975b-7e3baf922436" containerID="b5bb04c010728c244b89cd21ab4877c3c2544a4f2b0fc11c6b0ac82deea7e9ba" exitCode=0 Nov 22 14:45:34 crc kubenswrapper[4914]: I1122 14:45:34.042119 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zbdng" event={"ID":"9a4c03fc-56f1-4e81-975b-7e3baf922436","Type":"ContainerDied","Data":"b5bb04c010728c244b89cd21ab4877c3c2544a4f2b0fc11c6b0ac82deea7e9ba"} Nov 22 14:45:34 crc kubenswrapper[4914]: I1122 14:45:34.043750 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" event={"ID":"417645f9-c3c1-49c8-bcc4-84258eeb062d","Type":"ContainerStarted","Data":"19ee2fc85df36491a764da884c242dab024886b3c1f749c5a359a58e0a002aa4"} Nov 22 14:45:34 crc kubenswrapper[4914]: I1122 14:45:34.043804 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" event={"ID":"417645f9-c3c1-49c8-bcc4-84258eeb062d","Type":"ContainerStarted","Data":"56e9aa71fb08ca30002ab9f25d655c04631778ad2237186c7d1a1690acd48fff"} Nov 22 14:45:34 crc kubenswrapper[4914]: I1122 14:45:34.089701 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" podStartSLOduration=2.089683115 podStartE2EDuration="2.089683115s" podCreationTimestamp="2025-11-22 14:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:45:34.085181596 +0000 UTC m=+1099.893751297" watchObservedRunningTime="2025-11-22 14:45:34.089683115 +0000 UTC m=+1099.898252816" Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.052358 4914 generic.go:334] "Generic (PLEG): container finished" podID="417645f9-c3c1-49c8-bcc4-84258eeb062d" containerID="19ee2fc85df36491a764da884c242dab024886b3c1f749c5a359a58e0a002aa4" exitCode=0 Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.052433 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" event={"ID":"417645f9-c3c1-49c8-bcc4-84258eeb062d","Type":"ContainerDied","Data":"19ee2fc85df36491a764da884c242dab024886b3c1f749c5a359a58e0a002aa4"} Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.353701 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.365915 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4c03fc-56f1-4e81-975b-7e3baf922436-operator-scripts\") pod \"9a4c03fc-56f1-4e81-975b-7e3baf922436\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.366128 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctj8d\" (UniqueName: \"kubernetes.io/projected/9a4c03fc-56f1-4e81-975b-7e3baf922436-kube-api-access-ctj8d\") pod \"9a4c03fc-56f1-4e81-975b-7e3baf922436\" (UID: \"9a4c03fc-56f1-4e81-975b-7e3baf922436\") " Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.367828 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4c03fc-56f1-4e81-975b-7e3baf922436-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a4c03fc-56f1-4e81-975b-7e3baf922436" (UID: "9a4c03fc-56f1-4e81-975b-7e3baf922436"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.375953 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4c03fc-56f1-4e81-975b-7e3baf922436-kube-api-access-ctj8d" (OuterVolumeSpecName: "kube-api-access-ctj8d") pod "9a4c03fc-56f1-4e81-975b-7e3baf922436" (UID: "9a4c03fc-56f1-4e81-975b-7e3baf922436"). InnerVolumeSpecName "kube-api-access-ctj8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.467246 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a4c03fc-56f1-4e81-975b-7e3baf922436-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:35 crc kubenswrapper[4914]: I1122 14:45:35.467274 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctj8d\" (UniqueName: \"kubernetes.io/projected/9a4c03fc-56f1-4e81-975b-7e3baf922436-kube-api-access-ctj8d\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.064057 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zbdng" event={"ID":"9a4c03fc-56f1-4e81-975b-7e3baf922436","Type":"ContainerDied","Data":"fa51523087edb91e734665c820215d3f8ed2fd349a17a5fbbaca70a3b72dc284"} Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.064100 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa51523087edb91e734665c820215d3f8ed2fd349a17a5fbbaca70a3b72dc284" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.064098 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbdng" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.448662 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.584319 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48xjz\" (UniqueName: \"kubernetes.io/projected/417645f9-c3c1-49c8-bcc4-84258eeb062d-kube-api-access-48xjz\") pod \"417645f9-c3c1-49c8-bcc4-84258eeb062d\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.584484 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417645f9-c3c1-49c8-bcc4-84258eeb062d-operator-scripts\") pod \"417645f9-c3c1-49c8-bcc4-84258eeb062d\" (UID: \"417645f9-c3c1-49c8-bcc4-84258eeb062d\") " Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.585044 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/417645f9-c3c1-49c8-bcc4-84258eeb062d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "417645f9-c3c1-49c8-bcc4-84258eeb062d" (UID: "417645f9-c3c1-49c8-bcc4-84258eeb062d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.585698 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417645f9-c3c1-49c8-bcc4-84258eeb062d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.588778 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/417645f9-c3c1-49c8-bcc4-84258eeb062d-kube-api-access-48xjz" (OuterVolumeSpecName: "kube-api-access-48xjz") pod "417645f9-c3c1-49c8-bcc4-84258eeb062d" (UID: "417645f9-c3c1-49c8-bcc4-84258eeb062d"). InnerVolumeSpecName "kube-api-access-48xjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:36 crc kubenswrapper[4914]: I1122 14:45:36.686756 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48xjz\" (UniqueName: \"kubernetes.io/projected/417645f9-c3c1-49c8-bcc4-84258eeb062d-kube-api-access-48xjz\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.075352 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" event={"ID":"417645f9-c3c1-49c8-bcc4-84258eeb062d","Type":"ContainerDied","Data":"56e9aa71fb08ca30002ab9f25d655c04631778ad2237186c7d1a1690acd48fff"} Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.075394 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.075408 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56e9aa71fb08ca30002ab9f25d655c04631778ad2237186c7d1a1690acd48fff" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.570385 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-llch8"] Nov 22 14:45:37 crc kubenswrapper[4914]: E1122 14:45:37.570666 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4c03fc-56f1-4e81-975b-7e3baf922436" containerName="mariadb-database-create" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.570681 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4c03fc-56f1-4e81-975b-7e3baf922436" containerName="mariadb-database-create" Nov 22 14:45:37 crc kubenswrapper[4914]: E1122 14:45:37.570697 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417645f9-c3c1-49c8-bcc4-84258eeb062d" containerName="mariadb-account-create-update" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.570705 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="417645f9-c3c1-49c8-bcc4-84258eeb062d" containerName="mariadb-account-create-update" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.570872 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="417645f9-c3c1-49c8-bcc4-84258eeb062d" containerName="mariadb-account-create-update" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.570903 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4c03fc-56f1-4e81-975b-7e3baf922436" containerName="mariadb-database-create" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.571411 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.575079 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.575451 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.575772 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-n6sr5" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.586145 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-llch8"] Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.703438 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-config-data\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.703624 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z4sz\" (UniqueName: \"kubernetes.io/projected/3d0b543a-aa15-4f4f-b690-6dea37e042f7-kube-api-access-6z4sz\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.703715 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-combined-ca-bundle\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.703841 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-db-sync-config-data\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.805627 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-combined-ca-bundle\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.805730 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-db-sync-config-data\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.805814 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-config-data\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.806027 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z4sz\" (UniqueName: \"kubernetes.io/projected/3d0b543a-aa15-4f4f-b690-6dea37e042f7-kube-api-access-6z4sz\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.811327 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-db-sync-config-data\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.811998 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-config-data\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.814446 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-combined-ca-bundle\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.853436 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z4sz\" (UniqueName: \"kubernetes.io/projected/3d0b543a-aa15-4f4f-b690-6dea37e042f7-kube-api-access-6z4sz\") pod \"glance-db-sync-llch8\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:37 crc kubenswrapper[4914]: I1122 14:45:37.889263 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:38 crc kubenswrapper[4914]: I1122 14:45:38.324596 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-llch8"] Nov 22 14:45:38 crc kubenswrapper[4914]: W1122 14:45:38.330971 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d0b543a_aa15_4f4f_b690_6dea37e042f7.slice/crio-bfedec59d620a8f96d6ba68e07bf5d5d27fee4c86c361145e89aa8b7e5e05512 WatchSource:0}: Error finding container bfedec59d620a8f96d6ba68e07bf5d5d27fee4c86c361145e89aa8b7e5e05512: Status 404 returned error can't find the container with id bfedec59d620a8f96d6ba68e07bf5d5d27fee4c86c361145e89aa8b7e5e05512 Nov 22 14:45:39 crc kubenswrapper[4914]: I1122 14:45:39.093630 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-llch8" event={"ID":"3d0b543a-aa15-4f4f-b690-6dea37e042f7","Type":"ContainerStarted","Data":"903b9b43db8a3b106ce4025f53afbc4df3b0638aec7d72e3c68f54fada7a9f9c"} Nov 22 14:45:39 crc kubenswrapper[4914]: I1122 14:45:39.094376 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-llch8" event={"ID":"3d0b543a-aa15-4f4f-b690-6dea37e042f7","Type":"ContainerStarted","Data":"bfedec59d620a8f96d6ba68e07bf5d5d27fee4c86c361145e89aa8b7e5e05512"} Nov 22 14:45:39 crc kubenswrapper[4914]: I1122 14:45:39.112723 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-llch8" podStartSLOduration=2.112694454 podStartE2EDuration="2.112694454s" podCreationTimestamp="2025-11-22 14:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:45:39.109082797 +0000 UTC m=+1104.917652478" watchObservedRunningTime="2025-11-22 14:45:39.112694454 +0000 UTC m=+1104.921264125" Nov 22 14:45:42 crc kubenswrapper[4914]: I1122 14:45:42.119646 4914 generic.go:334] "Generic (PLEG): container finished" podID="3d0b543a-aa15-4f4f-b690-6dea37e042f7" containerID="903b9b43db8a3b106ce4025f53afbc4df3b0638aec7d72e3c68f54fada7a9f9c" exitCode=0 Nov 22 14:45:42 crc kubenswrapper[4914]: I1122 14:45:42.119735 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-llch8" event={"ID":"3d0b543a-aa15-4f4f-b690-6dea37e042f7","Type":"ContainerDied","Data":"903b9b43db8a3b106ce4025f53afbc4df3b0638aec7d72e3c68f54fada7a9f9c"} Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.436768 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.617817 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-config-data\") pod \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.617889 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z4sz\" (UniqueName: \"kubernetes.io/projected/3d0b543a-aa15-4f4f-b690-6dea37e042f7-kube-api-access-6z4sz\") pod \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.617944 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-db-sync-config-data\") pod \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.618003 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-combined-ca-bundle\") pod \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\" (UID: \"3d0b543a-aa15-4f4f-b690-6dea37e042f7\") " Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.622780 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3d0b543a-aa15-4f4f-b690-6dea37e042f7" (UID: "3d0b543a-aa15-4f4f-b690-6dea37e042f7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.624520 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d0b543a-aa15-4f4f-b690-6dea37e042f7-kube-api-access-6z4sz" (OuterVolumeSpecName: "kube-api-access-6z4sz") pod "3d0b543a-aa15-4f4f-b690-6dea37e042f7" (UID: "3d0b543a-aa15-4f4f-b690-6dea37e042f7"). InnerVolumeSpecName "kube-api-access-6z4sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.637975 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d0b543a-aa15-4f4f-b690-6dea37e042f7" (UID: "3d0b543a-aa15-4f4f-b690-6dea37e042f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.671117 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-config-data" (OuterVolumeSpecName: "config-data") pod "3d0b543a-aa15-4f4f-b690-6dea37e042f7" (UID: "3d0b543a-aa15-4f4f-b690-6dea37e042f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.720641 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.720677 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z4sz\" (UniqueName: \"kubernetes.io/projected/3d0b543a-aa15-4f4f-b690-6dea37e042f7-kube-api-access-6z4sz\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.720687 4914 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:43 crc kubenswrapper[4914]: I1122 14:45:43.720696 4914 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0b543a-aa15-4f4f-b690-6dea37e042f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.138338 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-llch8" event={"ID":"3d0b543a-aa15-4f4f-b690-6dea37e042f7","Type":"ContainerDied","Data":"bfedec59d620a8f96d6ba68e07bf5d5d27fee4c86c361145e89aa8b7e5e05512"} Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.138390 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfedec59d620a8f96d6ba68e07bf5d5d27fee4c86c361145e89aa8b7e5e05512" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.138461 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-llch8" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.339807 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:44 crc kubenswrapper[4914]: E1122 14:45:44.340397 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d0b543a-aa15-4f4f-b690-6dea37e042f7" containerName="glance-db-sync" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.340509 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d0b543a-aa15-4f4f-b690-6dea37e042f7" containerName="glance-db-sync" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.340794 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d0b543a-aa15-4f4f-b690-6dea37e042f7" containerName="glance-db-sync" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.341720 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.344412 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.344472 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.344485 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.344616 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-n6sr5" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.345241 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.345604 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.361149 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531104 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531302 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5hmc\" (UniqueName: \"kubernetes.io/projected/157f39ac-217e-4be9-a1eb-42c15381d96f-kube-api-access-c5hmc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531358 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531456 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-scripts\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531483 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-config-data\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531519 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531568 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-httpd-run\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531591 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-logs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.531695 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.632800 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.632899 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5hmc\" (UniqueName: \"kubernetes.io/projected/157f39ac-217e-4be9-a1eb-42c15381d96f-kube-api-access-c5hmc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.632945 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.632979 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-scripts\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.633002 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-config-data\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.633030 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.633804 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-httpd-run\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.633839 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-logs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.633881 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.634237 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.634268 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-logs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.634239 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-httpd-run\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.638731 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-scripts\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.639453 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.640645 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-config-data\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.641544 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.654606 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.654644 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5hmc\" (UniqueName: \"kubernetes.io/projected/157f39ac-217e-4be9-a1eb-42c15381d96f-kube-api-access-c5hmc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.661253 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:44 crc kubenswrapper[4914]: I1122 14:45:44.973422 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:45 crc kubenswrapper[4914]: I1122 14:45:45.219903 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:45 crc kubenswrapper[4914]: I1122 14:45:45.410281 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:46 crc kubenswrapper[4914]: I1122 14:45:46.158862 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"157f39ac-217e-4be9-a1eb-42c15381d96f","Type":"ContainerStarted","Data":"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d"} Nov 22 14:45:46 crc kubenswrapper[4914]: I1122 14:45:46.159186 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"157f39ac-217e-4be9-a1eb-42c15381d96f","Type":"ContainerStarted","Data":"e995dc2f96cc2a656efef60dd16bde00c82f25f140ed7c2911f1b765c5a83229"} Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.168696 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"157f39ac-217e-4be9-a1eb-42c15381d96f","Type":"ContainerStarted","Data":"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0"} Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.169013 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-httpd" containerID="cri-o://c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0" gracePeriod=30 Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.168855 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-log" containerID="cri-o://d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d" gracePeriod=30 Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.196972 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.196944079 podStartE2EDuration="3.196944079s" podCreationTimestamp="2025-11-22 14:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:45:47.192816819 +0000 UTC m=+1113.001386480" watchObservedRunningTime="2025-11-22 14:45:47.196944079 +0000 UTC m=+1113.005513780" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.673476 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816493 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-logs\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816659 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-config-data\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816713 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5hmc\" (UniqueName: \"kubernetes.io/projected/157f39ac-217e-4be9-a1eb-42c15381d96f-kube-api-access-c5hmc\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816758 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816806 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-combined-ca-bundle\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816881 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-scripts\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.816939 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-logs" (OuterVolumeSpecName: "logs") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.817028 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-httpd-run\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.817084 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-public-tls-certs\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.817118 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-internal-tls-certs\") pod \"157f39ac-217e-4be9-a1eb-42c15381d96f\" (UID: \"157f39ac-217e-4be9-a1eb-42c15381d96f\") " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.817466 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.817658 4914 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.817684 4914 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157f39ac-217e-4be9-a1eb-42c15381d96f-logs\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.822007 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.822116 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/157f39ac-217e-4be9-a1eb-42c15381d96f-kube-api-access-c5hmc" (OuterVolumeSpecName: "kube-api-access-c5hmc") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "kube-api-access-c5hmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.822130 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-scripts" (OuterVolumeSpecName: "scripts") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.839714 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.849857 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.853468 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-config-data" (OuterVolumeSpecName: "config-data") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.858413 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "157f39ac-217e-4be9-a1eb-42c15381d96f" (UID: "157f39ac-217e-4be9-a1eb-42c15381d96f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918362 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918390 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5hmc\" (UniqueName: \"kubernetes.io/projected/157f39ac-217e-4be9-a1eb-42c15381d96f-kube-api-access-c5hmc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918408 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918418 4914 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918427 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918434 4914 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.918444 4914 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/157f39ac-217e-4be9-a1eb-42c15381d96f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:47 crc kubenswrapper[4914]: I1122 14:45:47.931458 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.019208 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180719 4914 generic.go:334] "Generic (PLEG): container finished" podID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerID="c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0" exitCode=0 Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180754 4914 generic.go:334] "Generic (PLEG): container finished" podID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerID="d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d" exitCode=143 Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180778 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"157f39ac-217e-4be9-a1eb-42c15381d96f","Type":"ContainerDied","Data":"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0"} Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180817 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"157f39ac-217e-4be9-a1eb-42c15381d96f","Type":"ContainerDied","Data":"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d"} Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180819 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180860 4914 scope.go:117] "RemoveContainer" containerID="c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.180834 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"157f39ac-217e-4be9-a1eb-42c15381d96f","Type":"ContainerDied","Data":"e995dc2f96cc2a656efef60dd16bde00c82f25f140ed7c2911f1b765c5a83229"} Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.225783 4914 scope.go:117] "RemoveContainer" containerID="d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.228195 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.234691 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.252310 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:48 crc kubenswrapper[4914]: E1122 14:45:48.252773 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-httpd" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.252836 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-httpd" Nov 22 14:45:48 crc kubenswrapper[4914]: E1122 14:45:48.252896 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-log" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.253064 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-log" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.253239 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-log" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.253296 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" containerName="glance-httpd" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.253990 4914 scope.go:117] "RemoveContainer" containerID="c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.254176 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: E1122 14:45:48.255018 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0\": container with ID starting with c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0 not found: ID does not exist" containerID="c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.255059 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0"} err="failed to get container status \"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0\": rpc error: code = NotFound desc = could not find container \"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0\": container with ID starting with c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0 not found: ID does not exist" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.255087 4914 scope.go:117] "RemoveContainer" containerID="d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.257482 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.257708 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.257857 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.258087 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.258235 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-n6sr5" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.258371 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 22 14:45:48 crc kubenswrapper[4914]: E1122 14:45:48.259320 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d\": container with ID starting with d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d not found: ID does not exist" containerID="d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.259354 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d"} err="failed to get container status \"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d\": rpc error: code = NotFound desc = could not find container \"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d\": container with ID starting with d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d not found: ID does not exist" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.259383 4914 scope.go:117] "RemoveContainer" containerID="c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.265345 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0"} err="failed to get container status \"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0\": rpc error: code = NotFound desc = could not find container \"c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0\": container with ID starting with c381cc9d9c40ccf6528808de044873da8b0514d53557d812bcd4d4fac7b0acd0 not found: ID does not exist" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.265552 4914 scope.go:117] "RemoveContainer" containerID="d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.267007 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d"} err="failed to get container status \"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d\": rpc error: code = NotFound desc = could not find container \"d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d\": container with ID starting with d16239f37e01db43e9ac080d213bb567e29e6fff846ad5b3896c048fd1d2d49d not found: ID does not exist" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.271838 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.427172 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-config-data\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.427433 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.427574 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.427691 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.427787 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h5qp\" (UniqueName: \"kubernetes.io/projected/8d2c54c5-698a-4010-9ff4-233c33a78b0b-kube-api-access-7h5qp\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.427934 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.428037 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-httpd-run\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.428165 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-scripts\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.428266 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-logs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529420 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529469 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h5qp\" (UniqueName: \"kubernetes.io/projected/8d2c54c5-698a-4010-9ff4-233c33a78b0b-kube-api-access-7h5qp\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529501 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529529 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-httpd-run\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529593 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-scripts\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529618 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-logs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529657 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529682 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-config-data\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529717 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.529740 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.530288 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-logs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.530341 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-httpd-run\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.533369 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-scripts\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.533964 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.534004 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.534412 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.536267 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-config-data\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.556248 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h5qp\" (UniqueName: \"kubernetes.io/projected/8d2c54c5-698a-4010-9ff4-233c33a78b0b-kube-api-access-7h5qp\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.558613 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.633208 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:48 crc kubenswrapper[4914]: I1122 14:45:48.878502 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="157f39ac-217e-4be9-a1eb-42c15381d96f" path="/var/lib/kubelet/pods/157f39ac-217e-4be9-a1eb-42c15381d96f/volumes" Nov 22 14:45:49 crc kubenswrapper[4914]: I1122 14:45:49.041502 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:45:49 crc kubenswrapper[4914]: W1122 14:45:49.046000 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d2c54c5_698a_4010_9ff4_233c33a78b0b.slice/crio-1aa490bd646fd0a7c95c7405daed34c4002aebce3f178eadd7a3e33d515e41f2 WatchSource:0}: Error finding container 1aa490bd646fd0a7c95c7405daed34c4002aebce3f178eadd7a3e33d515e41f2: Status 404 returned error can't find the container with id 1aa490bd646fd0a7c95c7405daed34c4002aebce3f178eadd7a3e33d515e41f2 Nov 22 14:45:49 crc kubenswrapper[4914]: I1122 14:45:49.189734 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8d2c54c5-698a-4010-9ff4-233c33a78b0b","Type":"ContainerStarted","Data":"1aa490bd646fd0a7c95c7405daed34c4002aebce3f178eadd7a3e33d515e41f2"} Nov 22 14:45:50 crc kubenswrapper[4914]: I1122 14:45:50.202236 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8d2c54c5-698a-4010-9ff4-233c33a78b0b","Type":"ContainerStarted","Data":"e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4"} Nov 22 14:45:50 crc kubenswrapper[4914]: I1122 14:45:50.203215 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8d2c54c5-698a-4010-9ff4-233c33a78b0b","Type":"ContainerStarted","Data":"02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5"} Nov 22 14:45:50 crc kubenswrapper[4914]: I1122 14:45:50.225081 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.225060058 podStartE2EDuration="2.225060058s" podCreationTimestamp="2025-11-22 14:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:45:50.222137787 +0000 UTC m=+1116.030707468" watchObservedRunningTime="2025-11-22 14:45:50.225060058 +0000 UTC m=+1116.033629729" Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.043540 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.043965 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.044039 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.044981 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d84b3eee4b009a037e512f975ccca21eff513d79b47374207f1e9f4353ec16da"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.045080 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://d84b3eee4b009a037e512f975ccca21eff513d79b47374207f1e9f4353ec16da" gracePeriod=600 Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.228297 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="d84b3eee4b009a037e512f975ccca21eff513d79b47374207f1e9f4353ec16da" exitCode=0 Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.228393 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"d84b3eee4b009a037e512f975ccca21eff513d79b47374207f1e9f4353ec16da"} Nov 22 14:45:52 crc kubenswrapper[4914]: I1122 14:45:52.228451 4914 scope.go:117] "RemoveContainer" containerID="aa8dd9653a1cc07eb97e4563bba7df9b08941e911b0fd3c2ed65154f5dab470f" Nov 22 14:45:53 crc kubenswrapper[4914]: I1122 14:45:53.243945 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"d5e4a7a3ef4fbb39fd70fdad9ed4feaf8141c2c1da7a6c3705992baaafbc33f6"} Nov 22 14:45:58 crc kubenswrapper[4914]: I1122 14:45:58.634167 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:58 crc kubenswrapper[4914]: I1122 14:45:58.634767 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:58 crc kubenswrapper[4914]: I1122 14:45:58.660512 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:58 crc kubenswrapper[4914]: I1122 14:45:58.688135 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:59 crc kubenswrapper[4914]: I1122 14:45:59.316426 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:45:59 crc kubenswrapper[4914]: I1122 14:45:59.316481 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:46:01 crc kubenswrapper[4914]: I1122 14:46:01.218285 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:46:01 crc kubenswrapper[4914]: I1122 14:46:01.274289 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:46:02 crc kubenswrapper[4914]: I1122 14:46:02.900143 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-llch8"] Nov 22 14:46:02 crc kubenswrapper[4914]: I1122 14:46:02.907631 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-llch8"] Nov 22 14:46:02 crc kubenswrapper[4914]: I1122 14:46:02.971822 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancea3b4-account-delete-qfdtn"] Nov 22 14:46:02 crc kubenswrapper[4914]: I1122 14:46:02.972845 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:02 crc kubenswrapper[4914]: I1122 14:46:02.993824 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancea3b4-account-delete-qfdtn"] Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.063350 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/469f3b53-f112-40fb-a212-8771155eab66-operator-scripts\") pod \"glancea3b4-account-delete-qfdtn\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.063565 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx5h7\" (UniqueName: \"kubernetes.io/projected/469f3b53-f112-40fb-a212-8771155eab66-kube-api-access-fx5h7\") pod \"glancea3b4-account-delete-qfdtn\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.081985 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.164301 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/469f3b53-f112-40fb-a212-8771155eab66-operator-scripts\") pod \"glancea3b4-account-delete-qfdtn\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.164399 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx5h7\" (UniqueName: \"kubernetes.io/projected/469f3b53-f112-40fb-a212-8771155eab66-kube-api-access-fx5h7\") pod \"glancea3b4-account-delete-qfdtn\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.165264 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/469f3b53-f112-40fb-a212-8771155eab66-operator-scripts\") pod \"glancea3b4-account-delete-qfdtn\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.204765 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx5h7\" (UniqueName: \"kubernetes.io/projected/469f3b53-f112-40fb-a212-8771155eab66-kube-api-access-fx5h7\") pod \"glancea3b4-account-delete-qfdtn\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.289234 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.355381 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-log" containerID="cri-o://02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5" gracePeriod=30 Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.355678 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-httpd" containerID="cri-o://e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4" gracePeriod=30 Nov 22 14:46:03 crc kubenswrapper[4914]: I1122 14:46:03.554491 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancea3b4-account-delete-qfdtn"] Nov 22 14:46:03 crc kubenswrapper[4914]: W1122 14:46:03.557733 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod469f3b53_f112_40fb_a212_8771155eab66.slice/crio-1b50b8ae64864a5e5eab1110a3f779436bfbf209f9ef764ed683ae2a2428cb40 WatchSource:0}: Error finding container 1b50b8ae64864a5e5eab1110a3f779436bfbf209f9ef764ed683ae2a2428cb40: Status 404 returned error can't find the container with id 1b50b8ae64864a5e5eab1110a3f779436bfbf209f9ef764ed683ae2a2428cb40 Nov 22 14:46:04 crc kubenswrapper[4914]: I1122 14:46:04.364350 4914 generic.go:334] "Generic (PLEG): container finished" podID="469f3b53-f112-40fb-a212-8771155eab66" containerID="86133d606683f905138091f990f4c074fa56016a969c3824069e36917953961a" exitCode=0 Nov 22 14:46:04 crc kubenswrapper[4914]: I1122 14:46:04.364434 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" event={"ID":"469f3b53-f112-40fb-a212-8771155eab66","Type":"ContainerDied","Data":"86133d606683f905138091f990f4c074fa56016a969c3824069e36917953961a"} Nov 22 14:46:04 crc kubenswrapper[4914]: I1122 14:46:04.364462 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" event={"ID":"469f3b53-f112-40fb-a212-8771155eab66","Type":"ContainerStarted","Data":"1b50b8ae64864a5e5eab1110a3f779436bfbf209f9ef764ed683ae2a2428cb40"} Nov 22 14:46:04 crc kubenswrapper[4914]: I1122 14:46:04.366278 4914 generic.go:334] "Generic (PLEG): container finished" podID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerID="02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5" exitCode=143 Nov 22 14:46:04 crc kubenswrapper[4914]: I1122 14:46:04.366326 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8d2c54c5-698a-4010-9ff4-233c33a78b0b","Type":"ContainerDied","Data":"02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5"} Nov 22 14:46:04 crc kubenswrapper[4914]: I1122 14:46:04.884824 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d0b543a-aa15-4f4f-b690-6dea37e042f7" path="/var/lib/kubelet/pods/3d0b543a-aa15-4f4f-b690-6dea37e042f7/volumes" Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.685044 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.802840 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/469f3b53-f112-40fb-a212-8771155eab66-operator-scripts\") pod \"469f3b53-f112-40fb-a212-8771155eab66\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.802970 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx5h7\" (UniqueName: \"kubernetes.io/projected/469f3b53-f112-40fb-a212-8771155eab66-kube-api-access-fx5h7\") pod \"469f3b53-f112-40fb-a212-8771155eab66\" (UID: \"469f3b53-f112-40fb-a212-8771155eab66\") " Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.803910 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/469f3b53-f112-40fb-a212-8771155eab66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "469f3b53-f112-40fb-a212-8771155eab66" (UID: "469f3b53-f112-40fb-a212-8771155eab66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.808115 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469f3b53-f112-40fb-a212-8771155eab66-kube-api-access-fx5h7" (OuterVolumeSpecName: "kube-api-access-fx5h7") pod "469f3b53-f112-40fb-a212-8771155eab66" (UID: "469f3b53-f112-40fb-a212-8771155eab66"). InnerVolumeSpecName "kube-api-access-fx5h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.905399 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx5h7\" (UniqueName: \"kubernetes.io/projected/469f3b53-f112-40fb-a212-8771155eab66-kube-api-access-fx5h7\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:05 crc kubenswrapper[4914]: I1122 14:46:05.905718 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/469f3b53-f112-40fb-a212-8771155eab66-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:06 crc kubenswrapper[4914]: I1122 14:46:06.381815 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" event={"ID":"469f3b53-f112-40fb-a212-8771155eab66","Type":"ContainerDied","Data":"1b50b8ae64864a5e5eab1110a3f779436bfbf209f9ef764ed683ae2a2428cb40"} Nov 22 14:46:06 crc kubenswrapper[4914]: I1122 14:46:06.381857 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b50b8ae64864a5e5eab1110a3f779436bfbf209f9ef764ed683ae2a2428cb40" Nov 22 14:46:06 crc kubenswrapper[4914]: I1122 14:46:06.382134 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea3b4-account-delete-qfdtn" Nov 22 14:46:06 crc kubenswrapper[4914]: I1122 14:46:06.904395 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.024534 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h5qp\" (UniqueName: \"kubernetes.io/projected/8d2c54c5-698a-4010-9ff4-233c33a78b0b-kube-api-access-7h5qp\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.024893 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-scripts\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.024960 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-httpd-run\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.024993 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-combined-ca-bundle\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025015 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025038 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-config-data\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025112 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-internal-tls-certs\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025156 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025215 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-logs\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025820 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.025933 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-logs" (OuterVolumeSpecName: "logs") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.029640 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.030847 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2c54c5-698a-4010-9ff4-233c33a78b0b-kube-api-access-7h5qp" (OuterVolumeSpecName: "kube-api-access-7h5qp") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "kube-api-access-7h5qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.031851 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-scripts" (OuterVolumeSpecName: "scripts") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.070138 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.070709 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: E1122 14:46:07.083552 4914 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs podName:8d2c54c5-698a-4010-9ff4-233c33a78b0b nodeName:}" failed. No retries permitted until 2025-11-22 14:46:07.58351671 +0000 UTC m=+1133.392086371 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "public-tls-certs" (UniqueName: "kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b") : error deleting /var/lib/kubelet/pods/8d2c54c5-698a-4010-9ff4-233c33a78b0b/volume-subpaths: remove /var/lib/kubelet/pods/8d2c54c5-698a-4010-9ff4-233c33a78b0b/volume-subpaths: no such file or directory Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.087134 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-config-data" (OuterVolumeSpecName: "config-data") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126438 4914 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126471 4914 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126501 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126511 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126519 4914 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126528 4914 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2c54c5-698a-4010-9ff4-233c33a78b0b-logs\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126536 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h5qp\" (UniqueName: \"kubernetes.io/projected/8d2c54c5-698a-4010-9ff4-233c33a78b0b-kube-api-access-7h5qp\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.126544 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.140202 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.229990 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.389793 4914 generic.go:334] "Generic (PLEG): container finished" podID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerID="e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4" exitCode=0 Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.390044 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8d2c54c5-698a-4010-9ff4-233c33a78b0b","Type":"ContainerDied","Data":"e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4"} Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.390074 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8d2c54c5-698a-4010-9ff4-233c33a78b0b","Type":"ContainerDied","Data":"1aa490bd646fd0a7c95c7405daed34c4002aebce3f178eadd7a3e33d515e41f2"} Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.390093 4914 scope.go:117] "RemoveContainer" containerID="e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.390206 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.418933 4914 scope.go:117] "RemoveContainer" containerID="02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.441583 4914 scope.go:117] "RemoveContainer" containerID="e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4" Nov 22 14:46:07 crc kubenswrapper[4914]: E1122 14:46:07.445295 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4\": container with ID starting with e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4 not found: ID does not exist" containerID="e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.445337 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4"} err="failed to get container status \"e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4\": rpc error: code = NotFound desc = could not find container \"e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4\": container with ID starting with e21166be86bf777f5315180bf01eb0364042fc537bc867d714818ec60a549ac4 not found: ID does not exist" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.445369 4914 scope.go:117] "RemoveContainer" containerID="02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5" Nov 22 14:46:07 crc kubenswrapper[4914]: E1122 14:46:07.445956 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5\": container with ID starting with 02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5 not found: ID does not exist" containerID="02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.446009 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5"} err="failed to get container status \"02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5\": rpc error: code = NotFound desc = could not find container \"02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5\": container with ID starting with 02da0486a01c1b2f5c0f1bad18a5846236f0794710b8c4bd7a4fce99db5de2f5 not found: ID does not exist" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.634643 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs\") pod \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\" (UID: \"8d2c54c5-698a-4010-9ff4-233c33a78b0b\") " Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.643146 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8d2c54c5-698a-4010-9ff4-233c33a78b0b" (UID: "8d2c54c5-698a-4010-9ff4-233c33a78b0b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.720153 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.724377 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 22 14:46:07 crc kubenswrapper[4914]: I1122 14:46:07.736110 4914 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2c54c5-698a-4010-9ff4-233c33a78b0b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.002744 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbdng"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.007979 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbdng"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.013593 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancea3b4-account-delete-qfdtn"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.019064 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancea3b4-account-delete-qfdtn"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.023196 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.027012 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-a3b4-account-create-update-qg5gk"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.751254 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-k98mx"] Nov 22 14:46:08 crc kubenswrapper[4914]: E1122 14:46:08.751802 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-log" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.751814 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-log" Nov 22 14:46:08 crc kubenswrapper[4914]: E1122 14:46:08.751827 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469f3b53-f112-40fb-a212-8771155eab66" containerName="mariadb-account-delete" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.751834 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="469f3b53-f112-40fb-a212-8771155eab66" containerName="mariadb-account-delete" Nov 22 14:46:08 crc kubenswrapper[4914]: E1122 14:46:08.751853 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-httpd" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.751861 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-httpd" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.752151 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-log" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.752166 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="469f3b53-f112-40fb-a212-8771155eab66" containerName="mariadb-account-delete" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.752175 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" containerName="glance-httpd" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.752589 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.778743 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-9d43-account-create-update-wmdvf"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.779641 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.798102 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.798640 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-k98mx"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.857810 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9d43-account-create-update-wmdvf"] Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.858760 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxlp5\" (UniqueName: \"kubernetes.io/projected/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-kube-api-access-jxlp5\") pod \"glance-db-create-k98mx\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.858897 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lzgd\" (UniqueName: \"kubernetes.io/projected/fd247cef-eddf-4ee0-b170-e0b530b2dc48-kube-api-access-6lzgd\") pod \"glance-9d43-account-create-update-wmdvf\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.859029 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd247cef-eddf-4ee0-b170-e0b530b2dc48-operator-scripts\") pod \"glance-9d43-account-create-update-wmdvf\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.859094 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-operator-scripts\") pod \"glance-db-create-k98mx\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.895657 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="417645f9-c3c1-49c8-bcc4-84258eeb062d" path="/var/lib/kubelet/pods/417645f9-c3c1-49c8-bcc4-84258eeb062d/volumes" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.896363 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="469f3b53-f112-40fb-a212-8771155eab66" path="/var/lib/kubelet/pods/469f3b53-f112-40fb-a212-8771155eab66/volumes" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.896967 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2c54c5-698a-4010-9ff4-233c33a78b0b" path="/var/lib/kubelet/pods/8d2c54c5-698a-4010-9ff4-233c33a78b0b/volumes" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.898000 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4c03fc-56f1-4e81-975b-7e3baf922436" path="/var/lib/kubelet/pods/9a4c03fc-56f1-4e81-975b-7e3baf922436/volumes" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.960260 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lzgd\" (UniqueName: \"kubernetes.io/projected/fd247cef-eddf-4ee0-b170-e0b530b2dc48-kube-api-access-6lzgd\") pod \"glance-9d43-account-create-update-wmdvf\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.960373 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd247cef-eddf-4ee0-b170-e0b530b2dc48-operator-scripts\") pod \"glance-9d43-account-create-update-wmdvf\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.960402 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-operator-scripts\") pod \"glance-db-create-k98mx\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.960467 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxlp5\" (UniqueName: \"kubernetes.io/projected/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-kube-api-access-jxlp5\") pod \"glance-db-create-k98mx\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.961188 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd247cef-eddf-4ee0-b170-e0b530b2dc48-operator-scripts\") pod \"glance-9d43-account-create-update-wmdvf\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.961221 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-operator-scripts\") pod \"glance-db-create-k98mx\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.977656 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lzgd\" (UniqueName: \"kubernetes.io/projected/fd247cef-eddf-4ee0-b170-e0b530b2dc48-kube-api-access-6lzgd\") pod \"glance-9d43-account-create-update-wmdvf\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:08 crc kubenswrapper[4914]: I1122 14:46:08.987067 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxlp5\" (UniqueName: \"kubernetes.io/projected/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-kube-api-access-jxlp5\") pod \"glance-db-create-k98mx\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:09 crc kubenswrapper[4914]: I1122 14:46:09.081146 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:09 crc kubenswrapper[4914]: I1122 14:46:09.118321 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:09 crc kubenswrapper[4914]: I1122 14:46:09.508519 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-k98mx"] Nov 22 14:46:09 crc kubenswrapper[4914]: I1122 14:46:09.575769 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9d43-account-create-update-wmdvf"] Nov 22 14:46:09 crc kubenswrapper[4914]: W1122 14:46:09.580018 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd247cef_eddf_4ee0_b170_e0b530b2dc48.slice/crio-c044d80073caa8b6a0edad050040732c0a1e41bc9b948da14ffda211bf0f7876 WatchSource:0}: Error finding container c044d80073caa8b6a0edad050040732c0a1e41bc9b948da14ffda211bf0f7876: Status 404 returned error can't find the container with id c044d80073caa8b6a0edad050040732c0a1e41bc9b948da14ffda211bf0f7876 Nov 22 14:46:10 crc kubenswrapper[4914]: I1122 14:46:10.419500 4914 generic.go:334] "Generic (PLEG): container finished" podID="fd247cef-eddf-4ee0-b170-e0b530b2dc48" containerID="222158e92ce36c73b700992714d8f368e86766d28c2539a4a7e1886b1ecd71f4" exitCode=0 Nov 22 14:46:10 crc kubenswrapper[4914]: I1122 14:46:10.419575 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" event={"ID":"fd247cef-eddf-4ee0-b170-e0b530b2dc48","Type":"ContainerDied","Data":"222158e92ce36c73b700992714d8f368e86766d28c2539a4a7e1886b1ecd71f4"} Nov 22 14:46:10 crc kubenswrapper[4914]: I1122 14:46:10.419809 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" event={"ID":"fd247cef-eddf-4ee0-b170-e0b530b2dc48","Type":"ContainerStarted","Data":"c044d80073caa8b6a0edad050040732c0a1e41bc9b948da14ffda211bf0f7876"} Nov 22 14:46:10 crc kubenswrapper[4914]: I1122 14:46:10.422243 4914 generic.go:334] "Generic (PLEG): container finished" podID="a41c5010-2b3a-4229-bc2a-b1b6a09255e4" containerID="bcc3d47142545eb7dccb1d5341025fe89b2e6772130d3f6a8d7cfed8226b0946" exitCode=0 Nov 22 14:46:10 crc kubenswrapper[4914]: I1122 14:46:10.422272 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-k98mx" event={"ID":"a41c5010-2b3a-4229-bc2a-b1b6a09255e4","Type":"ContainerDied","Data":"bcc3d47142545eb7dccb1d5341025fe89b2e6772130d3f6a8d7cfed8226b0946"} Nov 22 14:46:10 crc kubenswrapper[4914]: I1122 14:46:10.422311 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-k98mx" event={"ID":"a41c5010-2b3a-4229-bc2a-b1b6a09255e4","Type":"ContainerStarted","Data":"4c7089150ae5d021b26e6b431e37e8841bfedc0c5766d49b2668dc7d5b63c071"} Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.759312 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.800677 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxlp5\" (UniqueName: \"kubernetes.io/projected/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-kube-api-access-jxlp5\") pod \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.800764 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-operator-scripts\") pod \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\" (UID: \"a41c5010-2b3a-4229-bc2a-b1b6a09255e4\") " Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.802622 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a41c5010-2b3a-4229-bc2a-b1b6a09255e4" (UID: "a41c5010-2b3a-4229-bc2a-b1b6a09255e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.806000 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-kube-api-access-jxlp5" (OuterVolumeSpecName: "kube-api-access-jxlp5") pod "a41c5010-2b3a-4229-bc2a-b1b6a09255e4" (UID: "a41c5010-2b3a-4229-bc2a-b1b6a09255e4"). InnerVolumeSpecName "kube-api-access-jxlp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.850437 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.902417 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd247cef-eddf-4ee0-b170-e0b530b2dc48-operator-scripts\") pod \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.902497 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lzgd\" (UniqueName: \"kubernetes.io/projected/fd247cef-eddf-4ee0-b170-e0b530b2dc48-kube-api-access-6lzgd\") pod \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\" (UID: \"fd247cef-eddf-4ee0-b170-e0b530b2dc48\") " Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.902815 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxlp5\" (UniqueName: \"kubernetes.io/projected/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-kube-api-access-jxlp5\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.902830 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41c5010-2b3a-4229-bc2a-b1b6a09255e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.903041 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd247cef-eddf-4ee0-b170-e0b530b2dc48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd247cef-eddf-4ee0-b170-e0b530b2dc48" (UID: "fd247cef-eddf-4ee0-b170-e0b530b2dc48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 14:46:11 crc kubenswrapper[4914]: I1122 14:46:11.905833 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd247cef-eddf-4ee0-b170-e0b530b2dc48-kube-api-access-6lzgd" (OuterVolumeSpecName: "kube-api-access-6lzgd") pod "fd247cef-eddf-4ee0-b170-e0b530b2dc48" (UID: "fd247cef-eddf-4ee0-b170-e0b530b2dc48"). InnerVolumeSpecName "kube-api-access-6lzgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.004664 4914 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd247cef-eddf-4ee0-b170-e0b530b2dc48-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.004695 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lzgd\" (UniqueName: \"kubernetes.io/projected/fd247cef-eddf-4ee0-b170-e0b530b2dc48-kube-api-access-6lzgd\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.441369 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-k98mx" Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.441913 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-k98mx" event={"ID":"a41c5010-2b3a-4229-bc2a-b1b6a09255e4","Type":"ContainerDied","Data":"4c7089150ae5d021b26e6b431e37e8841bfedc0c5766d49b2668dc7d5b63c071"} Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.442018 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c7089150ae5d021b26e6b431e37e8841bfedc0c5766d49b2668dc7d5b63c071" Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.443352 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" event={"ID":"fd247cef-eddf-4ee0-b170-e0b530b2dc48","Type":"ContainerDied","Data":"c044d80073caa8b6a0edad050040732c0a1e41bc9b948da14ffda211bf0f7876"} Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.443386 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c044d80073caa8b6a0edad050040732c0a1e41bc9b948da14ffda211bf0f7876" Nov 22 14:46:12 crc kubenswrapper[4914]: I1122 14:46:12.443418 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d43-account-create-update-wmdvf" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.075088 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-k792j"] Nov 22 14:46:14 crc kubenswrapper[4914]: E1122 14:46:14.075806 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41c5010-2b3a-4229-bc2a-b1b6a09255e4" containerName="mariadb-database-create" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.075827 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41c5010-2b3a-4229-bc2a-b1b6a09255e4" containerName="mariadb-database-create" Nov 22 14:46:14 crc kubenswrapper[4914]: E1122 14:46:14.075868 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd247cef-eddf-4ee0-b170-e0b530b2dc48" containerName="mariadb-account-create-update" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.075907 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd247cef-eddf-4ee0-b170-e0b530b2dc48" containerName="mariadb-account-create-update" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.076141 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd247cef-eddf-4ee0-b170-e0b530b2dc48" containerName="mariadb-account-create-update" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.076178 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="a41c5010-2b3a-4229-bc2a-b1b6a09255e4" containerName="mariadb-database-create" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.076988 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.079713 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.080129 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-z9r9z" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.086049 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-k792j"] Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.137216 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-config-data\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.137282 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-db-sync-config-data\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.137353 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7s28\" (UniqueName: \"kubernetes.io/projected/690d7575-7ec4-4487-a2ef-3ae7037c641d-kube-api-access-w7s28\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.238780 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-config-data\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.238863 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-db-sync-config-data\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.238947 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7s28\" (UniqueName: \"kubernetes.io/projected/690d7575-7ec4-4487-a2ef-3ae7037c641d-kube-api-access-w7s28\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.243712 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-db-sync-config-data\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.244342 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-config-data\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.276363 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7s28\" (UniqueName: \"kubernetes.io/projected/690d7575-7ec4-4487-a2ef-3ae7037c641d-kube-api-access-w7s28\") pod \"glance-db-sync-k792j\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.393209 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:14 crc kubenswrapper[4914]: I1122 14:46:14.934608 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-k792j"] Nov 22 14:46:15 crc kubenswrapper[4914]: I1122 14:46:15.472919 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-k792j" event={"ID":"690d7575-7ec4-4487-a2ef-3ae7037c641d","Type":"ContainerStarted","Data":"57f5890ed9ac52f12c2045f5d26dcbb7841dc2ae7cd02f93f3aa6cd054272a65"} Nov 22 14:46:16 crc kubenswrapper[4914]: I1122 14:46:16.485647 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-k792j" event={"ID":"690d7575-7ec4-4487-a2ef-3ae7037c641d","Type":"ContainerStarted","Data":"d038c3aca1f7b73e826709fef7de00239b42f3f3ab5c1849134c8d601e03d7d1"} Nov 22 14:46:16 crc kubenswrapper[4914]: I1122 14:46:16.512518 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-k792j" podStartSLOduration=2.512492207 podStartE2EDuration="2.512492207s" podCreationTimestamp="2025-11-22 14:46:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:46:16.505459776 +0000 UTC m=+1142.314029437" watchObservedRunningTime="2025-11-22 14:46:16.512492207 +0000 UTC m=+1142.321061898" Nov 22 14:46:18 crc kubenswrapper[4914]: I1122 14:46:18.504191 4914 generic.go:334] "Generic (PLEG): container finished" podID="690d7575-7ec4-4487-a2ef-3ae7037c641d" containerID="d038c3aca1f7b73e826709fef7de00239b42f3f3ab5c1849134c8d601e03d7d1" exitCode=0 Nov 22 14:46:18 crc kubenswrapper[4914]: I1122 14:46:18.504254 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-k792j" event={"ID":"690d7575-7ec4-4487-a2ef-3ae7037c641d","Type":"ContainerDied","Data":"d038c3aca1f7b73e826709fef7de00239b42f3f3ab5c1849134c8d601e03d7d1"} Nov 22 14:46:19 crc kubenswrapper[4914]: I1122 14:46:19.840035 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:19 crc kubenswrapper[4914]: I1122 14:46:19.973908 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-config-data\") pod \"690d7575-7ec4-4487-a2ef-3ae7037c641d\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " Nov 22 14:46:19 crc kubenswrapper[4914]: I1122 14:46:19.974085 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-db-sync-config-data\") pod \"690d7575-7ec4-4487-a2ef-3ae7037c641d\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " Nov 22 14:46:19 crc kubenswrapper[4914]: I1122 14:46:19.974151 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7s28\" (UniqueName: \"kubernetes.io/projected/690d7575-7ec4-4487-a2ef-3ae7037c641d-kube-api-access-w7s28\") pod \"690d7575-7ec4-4487-a2ef-3ae7037c641d\" (UID: \"690d7575-7ec4-4487-a2ef-3ae7037c641d\") " Nov 22 14:46:19 crc kubenswrapper[4914]: I1122 14:46:19.981247 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690d7575-7ec4-4487-a2ef-3ae7037c641d-kube-api-access-w7s28" (OuterVolumeSpecName: "kube-api-access-w7s28") pod "690d7575-7ec4-4487-a2ef-3ae7037c641d" (UID: "690d7575-7ec4-4487-a2ef-3ae7037c641d"). InnerVolumeSpecName "kube-api-access-w7s28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:46:19 crc kubenswrapper[4914]: I1122 14:46:19.981251 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "690d7575-7ec4-4487-a2ef-3ae7037c641d" (UID: "690d7575-7ec4-4487-a2ef-3ae7037c641d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.016710 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-config-data" (OuterVolumeSpecName: "config-data") pod "690d7575-7ec4-4487-a2ef-3ae7037c641d" (UID: "690d7575-7ec4-4487-a2ef-3ae7037c641d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.076286 4914 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.076333 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7s28\" (UniqueName: \"kubernetes.io/projected/690d7575-7ec4-4487-a2ef-3ae7037c641d-kube-api-access-w7s28\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.076348 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690d7575-7ec4-4487-a2ef-3ae7037c641d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.525039 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-k792j" event={"ID":"690d7575-7ec4-4487-a2ef-3ae7037c641d","Type":"ContainerDied","Data":"57f5890ed9ac52f12c2045f5d26dcbb7841dc2ae7cd02f93f3aa6cd054272a65"} Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.525709 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57f5890ed9ac52f12c2045f5d26dcbb7841dc2ae7cd02f93f3aa6cd054272a65" Nov 22 14:46:20 crc kubenswrapper[4914]: I1122 14:46:20.525161 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-k792j" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.468686 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 22 14:46:21 crc kubenswrapper[4914]: E1122 14:46:21.469925 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690d7575-7ec4-4487-a2ef-3ae7037c641d" containerName="glance-db-sync" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.469996 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="690d7575-7ec4-4487-a2ef-3ae7037c641d" containerName="glance-db-sync" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.470191 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="690d7575-7ec4-4487-a2ef-3ae7037c641d" containerName="glance-db-sync" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.471168 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.475725 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-z9r9z" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.475808 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.476681 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.484456 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.596635 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.596702 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb97ee27-5cea-4df8-87ff-72f05c460b73-logs\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.596764 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-dev\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.596792 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.596910 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-run\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.596968 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597022 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597053 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597098 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597370 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfgh4\" (UniqueName: \"kubernetes.io/projected/fb97ee27-5cea-4df8-87ff-72f05c460b73-kube-api-access-dfgh4\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597523 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb97ee27-5cea-4df8-87ff-72f05c460b73-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597619 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb97ee27-5cea-4df8-87ff-72f05c460b73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597702 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-sys\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.597809 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb97ee27-5cea-4df8-87ff-72f05c460b73-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699201 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699267 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-run\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699304 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699335 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699361 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699389 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699411 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfgh4\" (UniqueName: \"kubernetes.io/projected/fb97ee27-5cea-4df8-87ff-72f05c460b73-kube-api-access-dfgh4\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699441 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb97ee27-5cea-4df8-87ff-72f05c460b73-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699455 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb97ee27-5cea-4df8-87ff-72f05c460b73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699473 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-sys\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699494 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb97ee27-5cea-4df8-87ff-72f05c460b73-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699543 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699558 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb97ee27-5cea-4df8-87ff-72f05c460b73-logs\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699578 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-dev\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699643 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-dev\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699831 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.699901 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-run\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700021 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-sys\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700031 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700235 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700303 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700314 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700246 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fb97ee27-5cea-4df8-87ff-72f05c460b73-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.700444 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb97ee27-5cea-4df8-87ff-72f05c460b73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.702472 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb97ee27-5cea-4df8-87ff-72f05c460b73-logs\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.706344 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb97ee27-5cea-4df8-87ff-72f05c460b73-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.706578 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb97ee27-5cea-4df8-87ff-72f05c460b73-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.729295 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfgh4\" (UniqueName: \"kubernetes.io/projected/fb97ee27-5cea-4df8-87ff-72f05c460b73-kube-api-access-dfgh4\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.738950 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.742553 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"fb97ee27-5cea-4df8-87ff-72f05c460b73\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.789093 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.981789 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.983596 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:21 crc kubenswrapper[4914]: I1122 14:46:21.987459 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:21.997794 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108626 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108668 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108691 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-dev\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108710 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9mt\" (UniqueName: \"kubernetes.io/projected/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-kube-api-access-lm9mt\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108800 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-sys\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108868 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.108945 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109006 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109036 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-run\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109081 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109109 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109152 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-logs\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109185 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.109233 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211272 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-logs\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211317 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211345 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211372 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211403 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211422 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211447 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211495 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-dev\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211523 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9mt\" (UniqueName: \"kubernetes.io/projected/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-kube-api-access-lm9mt\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211544 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-sys\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211594 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211627 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211653 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211678 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-run\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211703 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.211851 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-logs\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212182 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212489 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-dev\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212560 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212716 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212748 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-sys\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212782 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212807 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212899 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.212970 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-run\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.217473 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.217983 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.233426 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.235801 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.244455 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9mt\" (UniqueName: \"kubernetes.io/projected/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-kube-api-access-lm9mt\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.249230 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.306720 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.542836 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.574899 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"fb97ee27-5cea-4df8-87ff-72f05c460b73","Type":"ContainerStarted","Data":"48f8614bcc523cab65387b27fb59ecb39252a1986957963d894ad65fe1b05563"} Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.574952 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"fb97ee27-5cea-4df8-87ff-72f05c460b73","Type":"ContainerStarted","Data":"eb6844cdebf0291e20ba3675954ff9f53eb533db547e9469264e1b76382b619a"} Nov 22 14:46:22 crc kubenswrapper[4914]: I1122 14:46:22.780292 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.584014 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"fb97ee27-5cea-4df8-87ff-72f05c460b73","Type":"ContainerStarted","Data":"7f53fa0f79ca062ceb1d7e6ba947bab4a6cd6f61650dada2dfacad460cbd8dfe"} Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.584569 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"fb97ee27-5cea-4df8-87ff-72f05c460b73","Type":"ContainerStarted","Data":"6a6e29ce22873010c9f9e95c3263560a8430ce5637fb0df5c8de4a9c306ca2e5"} Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.587715 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerStarted","Data":"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c"} Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.587745 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerStarted","Data":"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397"} Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.587760 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerStarted","Data":"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855"} Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.587770 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerStarted","Data":"626156e11616d869c66ef28c53e5ff3765573aae163ade4da7bf8e463274f6a9"} Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.587862 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-log" containerID="cri-o://d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" gracePeriod=30 Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.588251 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-api" containerID="cri-o://7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" gracePeriod=30 Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.588306 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-httpd" containerID="cri-o://37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" gracePeriod=30 Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.609292 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.6092704810000003 podStartE2EDuration="2.609270481s" podCreationTimestamp="2025-11-22 14:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:46:23.608623426 +0000 UTC m=+1149.417193087" watchObservedRunningTime="2025-11-22 14:46:23.609270481 +0000 UTC m=+1149.417840142" Nov 22 14:46:23 crc kubenswrapper[4914]: I1122 14:46:23.637713 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.6376863249999998 podStartE2EDuration="3.637686325s" podCreationTimestamp="2025-11-22 14:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:46:23.632402806 +0000 UTC m=+1149.440972487" watchObservedRunningTime="2025-11-22 14:46:23.637686325 +0000 UTC m=+1149.446256026" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.037413 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.142738 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-iscsi\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.142801 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm9mt\" (UniqueName: \"kubernetes.io/projected/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-kube-api-access-lm9mt\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.142825 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-httpd-run\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.142894 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-logs\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.142910 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.142944 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-var-locks-brick\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143000 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-config-data\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143021 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-lib-modules\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143041 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-nvme\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143082 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143119 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143149 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-run\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143173 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-dev\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143184 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-sys\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143194 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143208 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-logs" (OuterVolumeSpecName: "logs") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143204 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-scripts\") pod \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\" (UID: \"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d\") " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143568 4914 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143593 4914 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143605 4914 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-logs\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143638 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-run" (OuterVolumeSpecName: "run") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143782 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-dev" (OuterVolumeSpecName: "dev") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143831 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-sys" (OuterVolumeSpecName: "sys") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143857 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143893 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.143917 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.153987 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-scripts" (OuterVolumeSpecName: "scripts") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.153988 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.154024 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.154041 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-kube-api-access-lm9mt" (OuterVolumeSpecName: "kube-api-access-lm9mt") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "kube-api-access-lm9mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.200428 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-config-data" (OuterVolumeSpecName: "config-data") pod "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" (UID: "b64afbb6-e20f-47e1-8c2a-633d2b94aa1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244415 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm9mt\" (UniqueName: \"kubernetes.io/projected/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-kube-api-access-lm9mt\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244441 4914 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244451 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244460 4914 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244470 4914 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244499 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244510 4914 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244520 4914 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-run\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244530 4914 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-dev\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244537 4914 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-sys\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.244545 4914 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d-scripts\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.257625 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.264439 4914 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.345837 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.345916 4914 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.597439 4914 generic.go:334] "Generic (PLEG): container finished" podID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerID="7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" exitCode=143 Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.597470 4914 generic.go:334] "Generic (PLEG): container finished" podID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerID="37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" exitCode=143 Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.597476 4914 generic.go:334] "Generic (PLEG): container finished" podID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerID="d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" exitCode=143 Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.598413 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.599144 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerDied","Data":"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c"} Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.599208 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerDied","Data":"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397"} Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.599220 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerDied","Data":"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855"} Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.599228 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b64afbb6-e20f-47e1-8c2a-633d2b94aa1d","Type":"ContainerDied","Data":"626156e11616d869c66ef28c53e5ff3765573aae163ade4da7bf8e463274f6a9"} Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.599246 4914 scope.go:117] "RemoveContainer" containerID="7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.627169 4914 scope.go:117] "RemoveContainer" containerID="37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.640095 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.645809 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.673073 4914 scope.go:117] "RemoveContainer" containerID="d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.674947 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:24 crc kubenswrapper[4914]: E1122 14:46:24.675960 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-api" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.675997 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-api" Nov 22 14:46:24 crc kubenswrapper[4914]: E1122 14:46:24.676037 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-httpd" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.676051 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-httpd" Nov 22 14:46:24 crc kubenswrapper[4914]: E1122 14:46:24.676096 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-log" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.676110 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-log" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.676345 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-api" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.676370 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-httpd" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.676407 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" containerName="glance-log" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.678477 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.681320 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.695776 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.714157 4914 scope.go:117] "RemoveContainer" containerID="7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" Nov 22 14:46:24 crc kubenswrapper[4914]: E1122 14:46:24.717227 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": container with ID starting with 7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c not found: ID does not exist" containerID="7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.717256 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c"} err="failed to get container status \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": rpc error: code = NotFound desc = could not find container \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": container with ID starting with 7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.717277 4914 scope.go:117] "RemoveContainer" containerID="37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" Nov 22 14:46:24 crc kubenswrapper[4914]: E1122 14:46:24.719528 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": container with ID starting with 37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397 not found: ID does not exist" containerID="37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.719555 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397"} err="failed to get container status \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": rpc error: code = NotFound desc = could not find container \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": container with ID starting with 37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397 not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.719582 4914 scope.go:117] "RemoveContainer" containerID="d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" Nov 22 14:46:24 crc kubenswrapper[4914]: E1122 14:46:24.720380 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": container with ID starting with d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855 not found: ID does not exist" containerID="d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.720437 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855"} err="failed to get container status \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": rpc error: code = NotFound desc = could not find container \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": container with ID starting with d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855 not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.720478 4914 scope.go:117] "RemoveContainer" containerID="7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.721089 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c"} err="failed to get container status \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": rpc error: code = NotFound desc = could not find container \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": container with ID starting with 7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.721118 4914 scope.go:117] "RemoveContainer" containerID="37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.721612 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397"} err="failed to get container status \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": rpc error: code = NotFound desc = could not find container \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": container with ID starting with 37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397 not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.721639 4914 scope.go:117] "RemoveContainer" containerID="d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.721897 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855"} err="failed to get container status \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": rpc error: code = NotFound desc = could not find container \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": container with ID starting with d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855 not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.721923 4914 scope.go:117] "RemoveContainer" containerID="7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.722169 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c"} err="failed to get container status \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": rpc error: code = NotFound desc = could not find container \"7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c\": container with ID starting with 7be16a5c9f49e3a340558e7c5ae5b6c094ceabbc6beb09ffd34d3858d1da793c not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.722193 4914 scope.go:117] "RemoveContainer" containerID="37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.722893 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397"} err="failed to get container status \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": rpc error: code = NotFound desc = could not find container \"37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397\": container with ID starting with 37c865868eaa14972505db04dc9bc6f048d176ca251cfa9c2c38dd8d5877e397 not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.722924 4914 scope.go:117] "RemoveContainer" containerID="d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.724200 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855"} err="failed to get container status \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": rpc error: code = NotFound desc = could not find container \"d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855\": container with ID starting with d9821d9e59274a02443d3f223d59144381b0db68f404a9cbff8a293aeb871855 not found: ID does not exist" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751486 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751526 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gr4j\" (UniqueName: \"kubernetes.io/projected/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-kube-api-access-6gr4j\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751556 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751572 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-sys\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751589 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751611 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751630 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-run\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751668 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751687 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751704 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-logs\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751721 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751747 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751760 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.751781 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-dev\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852748 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852786 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852806 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-logs\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852824 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852852 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852866 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852904 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-dev\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852932 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852949 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gr4j\" (UniqueName: \"kubernetes.io/projected/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-kube-api-access-6gr4j\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.852974 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-sys\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853002 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853019 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853039 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853056 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-run\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853133 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-run\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853590 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-dev\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853642 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-sys\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853645 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853675 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853778 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853821 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853860 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.853996 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.854100 4914 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.854146 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-logs\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.857829 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.864644 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.874841 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gr4j\" (UniqueName: \"kubernetes.io/projected/0e0e08d2-dd7c-457e-ae5a-a3b350cd8927-kube-api-access-6gr4j\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.878365 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b64afbb6-e20f-47e1-8c2a-633d2b94aa1d" path="/var/lib/kubelet/pods/b64afbb6-e20f-47e1-8c2a-633d2b94aa1d/volumes" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.881054 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:24 crc kubenswrapper[4914]: I1122 14:46:24.893223 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:25 crc kubenswrapper[4914]: I1122 14:46:25.006596 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:25 crc kubenswrapper[4914]: I1122 14:46:25.518675 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 22 14:46:25 crc kubenswrapper[4914]: W1122 14:46:25.525197 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e0e08d2_dd7c_457e_ae5a_a3b350cd8927.slice/crio-0912b936907be817fbd16af72b62b93a4f42afbabef74b65a4af42a1cbb1cb4b WatchSource:0}: Error finding container 0912b936907be817fbd16af72b62b93a4f42afbabef74b65a4af42a1cbb1cb4b: Status 404 returned error can't find the container with id 0912b936907be817fbd16af72b62b93a4f42afbabef74b65a4af42a1cbb1cb4b Nov 22 14:46:25 crc kubenswrapper[4914]: I1122 14:46:25.608177 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927","Type":"ContainerStarted","Data":"0912b936907be817fbd16af72b62b93a4f42afbabef74b65a4af42a1cbb1cb4b"} Nov 22 14:46:26 crc kubenswrapper[4914]: I1122 14:46:26.618692 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927","Type":"ContainerStarted","Data":"d49dcea2ca64f358fd5aef6dd00968ca0cfd0c2ab8fc10eca95715fb59f1b888"} Nov 22 14:46:26 crc kubenswrapper[4914]: I1122 14:46:26.619320 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927","Type":"ContainerStarted","Data":"9d326246197ed8d8f00e8107bd0dec3eaf6a06608217caa61f900b437ce4281c"} Nov 22 14:46:26 crc kubenswrapper[4914]: I1122 14:46:26.619331 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"0e0e08d2-dd7c-457e-ae5a-a3b350cd8927","Type":"ContainerStarted","Data":"f881c7e6e473b66b8ab4c821c6a7cfdee578739259a987ea11d9dc3b4409e8bf"} Nov 22 14:46:26 crc kubenswrapper[4914]: I1122 14:46:26.648964 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.6489432920000002 podStartE2EDuration="2.648943292s" podCreationTimestamp="2025-11-22 14:46:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:46:26.647712873 +0000 UTC m=+1152.456282534" watchObservedRunningTime="2025-11-22 14:46:26.648943292 +0000 UTC m=+1152.457512973" Nov 22 14:46:31 crc kubenswrapper[4914]: I1122 14:46:31.789991 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:31 crc kubenswrapper[4914]: I1122 14:46:31.790929 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:31 crc kubenswrapper[4914]: I1122 14:46:31.790954 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:31 crc kubenswrapper[4914]: I1122 14:46:31.821726 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:31 crc kubenswrapper[4914]: I1122 14:46:31.822119 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:31 crc kubenswrapper[4914]: I1122 14:46:31.840503 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:32 crc kubenswrapper[4914]: I1122 14:46:32.675608 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:32 crc kubenswrapper[4914]: I1122 14:46:32.675649 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:32 crc kubenswrapper[4914]: I1122 14:46:32.675661 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:32 crc kubenswrapper[4914]: I1122 14:46:32.700773 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:32 crc kubenswrapper[4914]: I1122 14:46:32.706501 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:32 crc kubenswrapper[4914]: I1122 14:46:32.711476 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.007060 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.007289 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.007395 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.031718 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.038535 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.039794 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.701253 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.701298 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.701320 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.720298 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.725771 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:46:35 crc kubenswrapper[4914]: I1122 14:46:35.729677 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 22 14:47:19 crc kubenswrapper[4914]: I1122 14:47:19.088288 4914 scope.go:117] "RemoveContainer" containerID="cb1154e9ba5368a5ec609fc0f99acd3d993bd6c5f200ee1e9fa7ee22767c8a6f" Nov 22 14:47:19 crc kubenswrapper[4914]: I1122 14:47:19.113162 4914 scope.go:117] "RemoveContainer" containerID="ddb74e899dd5da4394f9d1c5b5c73ca52bba1ffb7bb4e8ed7f391e8a9a685c33" Nov 22 14:47:19 crc kubenswrapper[4914]: I1122 14:47:19.132891 4914 scope.go:117] "RemoveContainer" containerID="e52426445593c76f45c978236249effd59184a649a3fbf8d0ffde2fe568b357e" Nov 22 14:47:52 crc kubenswrapper[4914]: I1122 14:47:52.042519 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:47:52 crc kubenswrapper[4914]: I1122 14:47:52.043139 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:48:22 crc kubenswrapper[4914]: I1122 14:48:22.043056 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:48:22 crc kubenswrapper[4914]: I1122 14:48:22.043668 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.043012 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.043819 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.043917 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.044911 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5e4a7a3ef4fbb39fd70fdad9ed4feaf8141c2c1da7a6c3705992baaafbc33f6"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.045016 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://d5e4a7a3ef4fbb39fd70fdad9ed4feaf8141c2c1da7a6c3705992baaafbc33f6" gracePeriod=600 Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.864232 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="d5e4a7a3ef4fbb39fd70fdad9ed4feaf8141c2c1da7a6c3705992baaafbc33f6" exitCode=0 Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.864252 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"d5e4a7a3ef4fbb39fd70fdad9ed4feaf8141c2c1da7a6c3705992baaafbc33f6"} Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.864850 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de"} Nov 22 14:48:52 crc kubenswrapper[4914]: I1122 14:48:52.864886 4914 scope.go:117] "RemoveContainer" containerID="d84b3eee4b009a037e512f975ccca21eff513d79b47374207f1e9f4353ec16da" Nov 22 14:50:19 crc kubenswrapper[4914]: I1122 14:50:19.228180 4914 scope.go:117] "RemoveContainer" containerID="e44a87848313ee28111e3807e717a0cb1cb6e1eeea0a1187bf19023f121f9c7c" Nov 22 14:50:19 crc kubenswrapper[4914]: I1122 14:50:19.257904 4914 scope.go:117] "RemoveContainer" containerID="86a95b11c221f8bf340a236501c9ef17e941e13345ace7f8d3f0f51eaa7fa7e9" Nov 22 14:50:33 crc kubenswrapper[4914]: I1122 14:50:33.792502 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-85mvn"] Nov 22 14:50:33 crc kubenswrapper[4914]: I1122 14:50:33.795626 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:33 crc kubenswrapper[4914]: I1122 14:50:33.820817 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-85mvn"] Nov 22 14:50:33 crc kubenswrapper[4914]: I1122 14:50:33.967611 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-utilities\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:33 crc kubenswrapper[4914]: I1122 14:50:33.967738 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-catalog-content\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:33 crc kubenswrapper[4914]: I1122 14:50:33.968337 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzwhc\" (UniqueName: \"kubernetes.io/projected/2a5e9fff-4bb1-437e-aea6-3160bc307153-kube-api-access-jzwhc\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.070899 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-utilities\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.071519 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-utilities\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.072484 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-catalog-content\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.073163 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzwhc\" (UniqueName: \"kubernetes.io/projected/2a5e9fff-4bb1-437e-aea6-3160bc307153-kube-api-access-jzwhc\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.073202 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-catalog-content\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.097021 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzwhc\" (UniqueName: \"kubernetes.io/projected/2a5e9fff-4bb1-437e-aea6-3160bc307153-kube-api-access-jzwhc\") pod \"community-operators-85mvn\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.129537 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.619762 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-85mvn"] Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.817261 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerStarted","Data":"a0caba0653544f0fea45da4624aabf347766f7c0ef5052ef645b7d8d089f57a8"} Nov 22 14:50:34 crc kubenswrapper[4914]: I1122 14:50:34.818356 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerStarted","Data":"1296e109155c748d8d085936d37908a47292481ccd1a852e26b3535bd5f4e5a7"} Nov 22 14:50:35 crc kubenswrapper[4914]: I1122 14:50:35.827431 4914 generic.go:334] "Generic (PLEG): container finished" podID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerID="a0caba0653544f0fea45da4624aabf347766f7c0ef5052ef645b7d8d089f57a8" exitCode=0 Nov 22 14:50:35 crc kubenswrapper[4914]: I1122 14:50:35.827551 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerDied","Data":"a0caba0653544f0fea45da4624aabf347766f7c0ef5052ef645b7d8d089f57a8"} Nov 22 14:50:35 crc kubenswrapper[4914]: I1122 14:50:35.830551 4914 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 14:50:36 crc kubenswrapper[4914]: I1122 14:50:36.841140 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerStarted","Data":"54bfa18f711895bc2682da48911c6db9df5eb7e0e7a60e92ac7fab87a0e2d46a"} Nov 22 14:50:37 crc kubenswrapper[4914]: I1122 14:50:37.852852 4914 generic.go:334] "Generic (PLEG): container finished" podID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerID="54bfa18f711895bc2682da48911c6db9df5eb7e0e7a60e92ac7fab87a0e2d46a" exitCode=0 Nov 22 14:50:37 crc kubenswrapper[4914]: I1122 14:50:37.852963 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerDied","Data":"54bfa18f711895bc2682da48911c6db9df5eb7e0e7a60e92ac7fab87a0e2d46a"} Nov 22 14:50:38 crc kubenswrapper[4914]: I1122 14:50:38.866340 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerStarted","Data":"c8fbf12ac0f8da10314e4d0e28f8da323ea0a7e6591fdd4c44a97d5f77a13a9e"} Nov 22 14:50:38 crc kubenswrapper[4914]: I1122 14:50:38.904022 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-85mvn" podStartSLOduration=3.447746344 podStartE2EDuration="5.903995498s" podCreationTimestamp="2025-11-22 14:50:33 +0000 UTC" firstStartedPulling="2025-11-22 14:50:35.830135356 +0000 UTC m=+1401.638705057" lastFinishedPulling="2025-11-22 14:50:38.28638454 +0000 UTC m=+1404.094954211" observedRunningTime="2025-11-22 14:50:38.903069325 +0000 UTC m=+1404.711639046" watchObservedRunningTime="2025-11-22 14:50:38.903995498 +0000 UTC m=+1404.712565199" Nov 22 14:50:44 crc kubenswrapper[4914]: I1122 14:50:44.131101 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:44 crc kubenswrapper[4914]: I1122 14:50:44.133222 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:44 crc kubenswrapper[4914]: I1122 14:50:44.206343 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:44 crc kubenswrapper[4914]: I1122 14:50:44.988053 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:45 crc kubenswrapper[4914]: I1122 14:50:45.033461 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-85mvn"] Nov 22 14:50:46 crc kubenswrapper[4914]: I1122 14:50:46.933315 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-85mvn" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="registry-server" containerID="cri-o://c8fbf12ac0f8da10314e4d0e28f8da323ea0a7e6591fdd4c44a97d5f77a13a9e" gracePeriod=2 Nov 22 14:50:47 crc kubenswrapper[4914]: I1122 14:50:47.949285 4914 generic.go:334] "Generic (PLEG): container finished" podID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerID="c8fbf12ac0f8da10314e4d0e28f8da323ea0a7e6591fdd4c44a97d5f77a13a9e" exitCode=0 Nov 22 14:50:47 crc kubenswrapper[4914]: I1122 14:50:47.949351 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerDied","Data":"c8fbf12ac0f8da10314e4d0e28f8da323ea0a7e6591fdd4c44a97d5f77a13a9e"} Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.082626 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.253345 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-catalog-content\") pod \"2a5e9fff-4bb1-437e-aea6-3160bc307153\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.253809 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-utilities\") pod \"2a5e9fff-4bb1-437e-aea6-3160bc307153\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.253853 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzwhc\" (UniqueName: \"kubernetes.io/projected/2a5e9fff-4bb1-437e-aea6-3160bc307153-kube-api-access-jzwhc\") pod \"2a5e9fff-4bb1-437e-aea6-3160bc307153\" (UID: \"2a5e9fff-4bb1-437e-aea6-3160bc307153\") " Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.255004 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-utilities" (OuterVolumeSpecName: "utilities") pod "2a5e9fff-4bb1-437e-aea6-3160bc307153" (UID: "2a5e9fff-4bb1-437e-aea6-3160bc307153"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.259900 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a5e9fff-4bb1-437e-aea6-3160bc307153-kube-api-access-jzwhc" (OuterVolumeSpecName: "kube-api-access-jzwhc") pod "2a5e9fff-4bb1-437e-aea6-3160bc307153" (UID: "2a5e9fff-4bb1-437e-aea6-3160bc307153"). InnerVolumeSpecName "kube-api-access-jzwhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.303374 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a5e9fff-4bb1-437e-aea6-3160bc307153" (UID: "2a5e9fff-4bb1-437e-aea6-3160bc307153"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.355513 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.355551 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a5e9fff-4bb1-437e-aea6-3160bc307153-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.355566 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzwhc\" (UniqueName: \"kubernetes.io/projected/2a5e9fff-4bb1-437e-aea6-3160bc307153-kube-api-access-jzwhc\") on node \"crc\" DevicePath \"\"" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.958815 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85mvn" event={"ID":"2a5e9fff-4bb1-437e-aea6-3160bc307153","Type":"ContainerDied","Data":"1296e109155c748d8d085936d37908a47292481ccd1a852e26b3535bd5f4e5a7"} Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.958877 4914 scope.go:117] "RemoveContainer" containerID="c8fbf12ac0f8da10314e4d0e28f8da323ea0a7e6591fdd4c44a97d5f77a13a9e" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.959031 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85mvn" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.987610 4914 scope.go:117] "RemoveContainer" containerID="54bfa18f711895bc2682da48911c6db9df5eb7e0e7a60e92ac7fab87a0e2d46a" Nov 22 14:50:48 crc kubenswrapper[4914]: I1122 14:50:48.994181 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-85mvn"] Nov 22 14:50:49 crc kubenswrapper[4914]: I1122 14:50:49.000752 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-85mvn"] Nov 22 14:50:49 crc kubenswrapper[4914]: I1122 14:50:49.012354 4914 scope.go:117] "RemoveContainer" containerID="a0caba0653544f0fea45da4624aabf347766f7c0ef5052ef645b7d8d089f57a8" Nov 22 14:50:50 crc kubenswrapper[4914]: I1122 14:50:50.876827 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" path="/var/lib/kubelet/pods/2a5e9fff-4bb1-437e-aea6-3160bc307153/volumes" Nov 22 14:50:52 crc kubenswrapper[4914]: I1122 14:50:52.042515 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:50:52 crc kubenswrapper[4914]: I1122 14:50:52.042598 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.663673 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-czkjh"] Nov 22 14:51:01 crc kubenswrapper[4914]: E1122 14:51:01.664608 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="registry-server" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.664624 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="registry-server" Nov 22 14:51:01 crc kubenswrapper[4914]: E1122 14:51:01.664660 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="extract-utilities" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.664668 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="extract-utilities" Nov 22 14:51:01 crc kubenswrapper[4914]: E1122 14:51:01.664688 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="extract-content" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.664695 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="extract-content" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.664868 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a5e9fff-4bb1-437e-aea6-3160bc307153" containerName="registry-server" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.666133 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.683681 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-utilities\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.683775 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6kms\" (UniqueName: \"kubernetes.io/projected/2c888010-856b-4919-a74d-7d7d7f5ebd5c-kube-api-access-n6kms\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.683803 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-catalog-content\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.688476 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czkjh"] Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.784663 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-utilities\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.784725 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6kms\" (UniqueName: \"kubernetes.io/projected/2c888010-856b-4919-a74d-7d7d7f5ebd5c-kube-api-access-n6kms\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.784752 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-catalog-content\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.785175 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-utilities\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.785355 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-catalog-content\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.810296 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6kms\" (UniqueName: \"kubernetes.io/projected/2c888010-856b-4919-a74d-7d7d7f5ebd5c-kube-api-access-n6kms\") pod \"redhat-operators-czkjh\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:01 crc kubenswrapper[4914]: I1122 14:51:01.984192 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:02 crc kubenswrapper[4914]: I1122 14:51:02.392216 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czkjh"] Nov 22 14:51:03 crc kubenswrapper[4914]: I1122 14:51:03.105965 4914 generic.go:334] "Generic (PLEG): container finished" podID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerID="95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963" exitCode=0 Nov 22 14:51:03 crc kubenswrapper[4914]: I1122 14:51:03.106085 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czkjh" event={"ID":"2c888010-856b-4919-a74d-7d7d7f5ebd5c","Type":"ContainerDied","Data":"95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963"} Nov 22 14:51:03 crc kubenswrapper[4914]: I1122 14:51:03.106262 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czkjh" event={"ID":"2c888010-856b-4919-a74d-7d7d7f5ebd5c","Type":"ContainerStarted","Data":"37cca7855a190c51aadd3834d51d7d3706b3028a9944a0f7cc5b08ee9b3833af"} Nov 22 14:51:05 crc kubenswrapper[4914]: I1122 14:51:05.124352 4914 generic.go:334] "Generic (PLEG): container finished" podID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerID="b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b" exitCode=0 Nov 22 14:51:05 crc kubenswrapper[4914]: I1122 14:51:05.124461 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czkjh" event={"ID":"2c888010-856b-4919-a74d-7d7d7f5ebd5c","Type":"ContainerDied","Data":"b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b"} Nov 22 14:51:06 crc kubenswrapper[4914]: I1122 14:51:06.133041 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czkjh" event={"ID":"2c888010-856b-4919-a74d-7d7d7f5ebd5c","Type":"ContainerStarted","Data":"97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1"} Nov 22 14:51:06 crc kubenswrapper[4914]: I1122 14:51:06.155500 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-czkjh" podStartSLOduration=2.726060409 podStartE2EDuration="5.155476445s" podCreationTimestamp="2025-11-22 14:51:01 +0000 UTC" firstStartedPulling="2025-11-22 14:51:03.10758462 +0000 UTC m=+1428.916154291" lastFinishedPulling="2025-11-22 14:51:05.537000646 +0000 UTC m=+1431.345570327" observedRunningTime="2025-11-22 14:51:06.151265232 +0000 UTC m=+1431.959834893" watchObservedRunningTime="2025-11-22 14:51:06.155476445 +0000 UTC m=+1431.964046116" Nov 22 14:51:11 crc kubenswrapper[4914]: I1122 14:51:11.985518 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:11 crc kubenswrapper[4914]: I1122 14:51:11.986093 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:13 crc kubenswrapper[4914]: I1122 14:51:13.042180 4914 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-czkjh" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="registry-server" probeResult="failure" output=< Nov 22 14:51:13 crc kubenswrapper[4914]: timeout: failed to connect service ":50051" within 1s Nov 22 14:51:13 crc kubenswrapper[4914]: > Nov 22 14:51:19 crc kubenswrapper[4914]: I1122 14:51:19.364295 4914 scope.go:117] "RemoveContainer" containerID="7dd998c7d54cddf21070c91b89f3f1bba5aa6ada3d742f23365ac28c55f663c1" Nov 22 14:51:22 crc kubenswrapper[4914]: I1122 14:51:22.042604 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:51:22 crc kubenswrapper[4914]: I1122 14:51:22.043407 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:51:22 crc kubenswrapper[4914]: I1122 14:51:22.068160 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:22 crc kubenswrapper[4914]: I1122 14:51:22.117519 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:22 crc kubenswrapper[4914]: I1122 14:51:22.306145 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-czkjh"] Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.278425 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-czkjh" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="registry-server" containerID="cri-o://97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1" gracePeriod=2 Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.799510 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.878206 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6kms\" (UniqueName: \"kubernetes.io/projected/2c888010-856b-4919-a74d-7d7d7f5ebd5c-kube-api-access-n6kms\") pod \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.878433 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-catalog-content\") pod \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.878525 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-utilities\") pod \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\" (UID: \"2c888010-856b-4919-a74d-7d7d7f5ebd5c\") " Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.879799 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-utilities" (OuterVolumeSpecName: "utilities") pod "2c888010-856b-4919-a74d-7d7d7f5ebd5c" (UID: "2c888010-856b-4919-a74d-7d7d7f5ebd5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:51:23 crc kubenswrapper[4914]: I1122 14:51:23.887091 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c888010-856b-4919-a74d-7d7d7f5ebd5c-kube-api-access-n6kms" (OuterVolumeSpecName: "kube-api-access-n6kms") pod "2c888010-856b-4919-a74d-7d7d7f5ebd5c" (UID: "2c888010-856b-4919-a74d-7d7d7f5ebd5c"). InnerVolumeSpecName "kube-api-access-n6kms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:23.980659 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6kms\" (UniqueName: \"kubernetes.io/projected/2c888010-856b-4919-a74d-7d7d7f5ebd5c-kube-api-access-n6kms\") on node \"crc\" DevicePath \"\"" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:23.980687 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.056396 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c888010-856b-4919-a74d-7d7d7f5ebd5c" (UID: "2c888010-856b-4919-a74d-7d7d7f5ebd5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.082813 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c888010-856b-4919-a74d-7d7d7f5ebd5c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.292054 4914 generic.go:334] "Generic (PLEG): container finished" podID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerID="97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1" exitCode=0 Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.292094 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czkjh" event={"ID":"2c888010-856b-4919-a74d-7d7d7f5ebd5c","Type":"ContainerDied","Data":"97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1"} Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.292134 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czkjh" event={"ID":"2c888010-856b-4919-a74d-7d7d7f5ebd5c","Type":"ContainerDied","Data":"37cca7855a190c51aadd3834d51d7d3706b3028a9944a0f7cc5b08ee9b3833af"} Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.292155 4914 scope.go:117] "RemoveContainer" containerID="97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.292162 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czkjh" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.323456 4914 scope.go:117] "RemoveContainer" containerID="b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.333345 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-czkjh"] Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.338333 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-czkjh"] Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.355069 4914 scope.go:117] "RemoveContainer" containerID="95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.393383 4914 scope.go:117] "RemoveContainer" containerID="97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1" Nov 22 14:51:24 crc kubenswrapper[4914]: E1122 14:51:24.393867 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1\": container with ID starting with 97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1 not found: ID does not exist" containerID="97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.393948 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1"} err="failed to get container status \"97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1\": rpc error: code = NotFound desc = could not find container \"97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1\": container with ID starting with 97d8c7146e8bba18b9c6cc8d7bae6d13aab0def90899c30303ed0fe14aa764c1 not found: ID does not exist" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.393981 4914 scope.go:117] "RemoveContainer" containerID="b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b" Nov 22 14:51:24 crc kubenswrapper[4914]: E1122 14:51:24.394481 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b\": container with ID starting with b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b not found: ID does not exist" containerID="b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.394509 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b"} err="failed to get container status \"b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b\": rpc error: code = NotFound desc = could not find container \"b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b\": container with ID starting with b828406ed587ca8096de3dcd8f1a505293b86eb3cb0d13dc58fc5a70613f1a6b not found: ID does not exist" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.394526 4914 scope.go:117] "RemoveContainer" containerID="95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963" Nov 22 14:51:24 crc kubenswrapper[4914]: E1122 14:51:24.394895 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963\": container with ID starting with 95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963 not found: ID does not exist" containerID="95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.394991 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963"} err="failed to get container status \"95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963\": rpc error: code = NotFound desc = could not find container \"95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963\": container with ID starting with 95070a759dcb234a212fd1011a1087f9682870ba241e85c12ba28399de6ec963 not found: ID does not exist" Nov 22 14:51:24 crc kubenswrapper[4914]: I1122 14:51:24.880504 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" path="/var/lib/kubelet/pods/2c888010-856b-4919-a74d-7d7d7f5ebd5c/volumes" Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.042128 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.042767 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.042844 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.043686 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.043752 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" gracePeriod=600 Nov 22 14:51:52 crc kubenswrapper[4914]: E1122 14:51:52.175122 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.517433 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" exitCode=0 Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.517477 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de"} Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.517517 4914 scope.go:117] "RemoveContainer" containerID="d5e4a7a3ef4fbb39fd70fdad9ed4feaf8141c2c1da7a6c3705992baaafbc33f6" Nov 22 14:51:52 crc kubenswrapper[4914]: I1122 14:51:52.518040 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:51:52 crc kubenswrapper[4914]: E1122 14:51:52.518261 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.672341 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wqfm2"] Nov 22 14:51:57 crc kubenswrapper[4914]: E1122 14:51:57.673080 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="extract-content" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.673102 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="extract-content" Nov 22 14:51:57 crc kubenswrapper[4914]: E1122 14:51:57.673144 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="extract-utilities" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.673156 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="extract-utilities" Nov 22 14:51:57 crc kubenswrapper[4914]: E1122 14:51:57.673201 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="registry-server" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.673214 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="registry-server" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.673440 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c888010-856b-4919-a74d-7d7d7f5ebd5c" containerName="registry-server" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.675068 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.722354 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqfm2"] Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.726390 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-catalog-content\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.726546 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bk9x\" (UniqueName: \"kubernetes.io/projected/10f05c4b-2473-4efd-b408-e9870dff116f-kube-api-access-6bk9x\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.726631 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-utilities\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.828232 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-utilities\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.828364 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-catalog-content\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.828448 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bk9x\" (UniqueName: \"kubernetes.io/projected/10f05c4b-2473-4efd-b408-e9870dff116f-kube-api-access-6bk9x\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.828894 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-utilities\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.828947 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-catalog-content\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:57 crc kubenswrapper[4914]: I1122 14:51:57.868424 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bk9x\" (UniqueName: \"kubernetes.io/projected/10f05c4b-2473-4efd-b408-e9870dff116f-kube-api-access-6bk9x\") pod \"certified-operators-wqfm2\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:58 crc kubenswrapper[4914]: I1122 14:51:58.001493 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:51:58 crc kubenswrapper[4914]: I1122 14:51:58.440107 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqfm2"] Nov 22 14:51:58 crc kubenswrapper[4914]: I1122 14:51:58.572903 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqfm2" event={"ID":"10f05c4b-2473-4efd-b408-e9870dff116f","Type":"ContainerStarted","Data":"2cd9e76da1cc52457585ee4d4ac058d5c606694d2e6690023bdc03c0ce8c8853"} Nov 22 14:51:59 crc kubenswrapper[4914]: I1122 14:51:59.581400 4914 generic.go:334] "Generic (PLEG): container finished" podID="10f05c4b-2473-4efd-b408-e9870dff116f" containerID="7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f" exitCode=0 Nov 22 14:51:59 crc kubenswrapper[4914]: I1122 14:51:59.581491 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqfm2" event={"ID":"10f05c4b-2473-4efd-b408-e9870dff116f","Type":"ContainerDied","Data":"7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f"} Nov 22 14:52:00 crc kubenswrapper[4914]: I1122 14:52:00.592265 4914 generic.go:334] "Generic (PLEG): container finished" podID="10f05c4b-2473-4efd-b408-e9870dff116f" containerID="92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd" exitCode=0 Nov 22 14:52:00 crc kubenswrapper[4914]: I1122 14:52:00.592368 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqfm2" event={"ID":"10f05c4b-2473-4efd-b408-e9870dff116f","Type":"ContainerDied","Data":"92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd"} Nov 22 14:52:01 crc kubenswrapper[4914]: I1122 14:52:01.601272 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqfm2" event={"ID":"10f05c4b-2473-4efd-b408-e9870dff116f","Type":"ContainerStarted","Data":"1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb"} Nov 22 14:52:01 crc kubenswrapper[4914]: I1122 14:52:01.620112 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wqfm2" podStartSLOduration=3.213289665 podStartE2EDuration="4.62009158s" podCreationTimestamp="2025-11-22 14:51:57 +0000 UTC" firstStartedPulling="2025-11-22 14:51:59.583240673 +0000 UTC m=+1485.391810334" lastFinishedPulling="2025-11-22 14:52:00.990042588 +0000 UTC m=+1486.798612249" observedRunningTime="2025-11-22 14:52:01.617531628 +0000 UTC m=+1487.426101319" watchObservedRunningTime="2025-11-22 14:52:01.62009158 +0000 UTC m=+1487.428661241" Nov 22 14:52:03 crc kubenswrapper[4914]: I1122 14:52:03.050994 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4qjxm"] Nov 22 14:52:03 crc kubenswrapper[4914]: I1122 14:52:03.058597 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h"] Nov 22 14:52:03 crc kubenswrapper[4914]: I1122 14:52:03.065152 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4qjxm"] Nov 22 14:52:03 crc kubenswrapper[4914]: I1122 14:52:03.071040 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-0bcc-account-create-update-qpb9h"] Nov 22 14:52:04 crc kubenswrapper[4914]: I1122 14:52:04.875068 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ba56179-4086-4862-8a37-cd9a4ccdf9bd" path="/var/lib/kubelet/pods/5ba56179-4086-4862-8a37-cd9a4ccdf9bd/volumes" Nov 22 14:52:04 crc kubenswrapper[4914]: I1122 14:52:04.875919 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d71cfef0-c0bc-48af-8449-3a902ce32e1f" path="/var/lib/kubelet/pods/d71cfef0-c0bc-48af-8449-3a902ce32e1f/volumes" Nov 22 14:52:05 crc kubenswrapper[4914]: I1122 14:52:05.868546 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:52:05 crc kubenswrapper[4914]: E1122 14:52:05.868826 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:52:08 crc kubenswrapper[4914]: I1122 14:52:08.002333 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:52:08 crc kubenswrapper[4914]: I1122 14:52:08.003750 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:52:08 crc kubenswrapper[4914]: I1122 14:52:08.059203 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:52:08 crc kubenswrapper[4914]: I1122 14:52:08.735642 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:52:08 crc kubenswrapper[4914]: I1122 14:52:08.790945 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqfm2"] Nov 22 14:52:10 crc kubenswrapper[4914]: I1122 14:52:10.696295 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wqfm2" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="registry-server" containerID="cri-o://1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb" gracePeriod=2 Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.074342 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.163274 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bk9x\" (UniqueName: \"kubernetes.io/projected/10f05c4b-2473-4efd-b408-e9870dff116f-kube-api-access-6bk9x\") pod \"10f05c4b-2473-4efd-b408-e9870dff116f\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.163527 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-utilities\") pod \"10f05c4b-2473-4efd-b408-e9870dff116f\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.163554 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-catalog-content\") pod \"10f05c4b-2473-4efd-b408-e9870dff116f\" (UID: \"10f05c4b-2473-4efd-b408-e9870dff116f\") " Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.164896 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-utilities" (OuterVolumeSpecName: "utilities") pod "10f05c4b-2473-4efd-b408-e9870dff116f" (UID: "10f05c4b-2473-4efd-b408-e9870dff116f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.169989 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f05c4b-2473-4efd-b408-e9870dff116f-kube-api-access-6bk9x" (OuterVolumeSpecName: "kube-api-access-6bk9x") pod "10f05c4b-2473-4efd-b408-e9870dff116f" (UID: "10f05c4b-2473-4efd-b408-e9870dff116f"). InnerVolumeSpecName "kube-api-access-6bk9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.265466 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.265515 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bk9x\" (UniqueName: \"kubernetes.io/projected/10f05c4b-2473-4efd-b408-e9870dff116f-kube-api-access-6bk9x\") on node \"crc\" DevicePath \"\"" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.704275 4914 generic.go:334] "Generic (PLEG): container finished" podID="10f05c4b-2473-4efd-b408-e9870dff116f" containerID="1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb" exitCode=0 Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.704328 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqfm2" event={"ID":"10f05c4b-2473-4efd-b408-e9870dff116f","Type":"ContainerDied","Data":"1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb"} Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.704333 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqfm2" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.704357 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqfm2" event={"ID":"10f05c4b-2473-4efd-b408-e9870dff116f","Type":"ContainerDied","Data":"2cd9e76da1cc52457585ee4d4ac058d5c606694d2e6690023bdc03c0ce8c8853"} Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.704380 4914 scope.go:117] "RemoveContainer" containerID="1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.719843 4914 scope.go:117] "RemoveContainer" containerID="92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.790792 4914 scope.go:117] "RemoveContainer" containerID="7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.812103 4914 scope.go:117] "RemoveContainer" containerID="1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb" Nov 22 14:52:11 crc kubenswrapper[4914]: E1122 14:52:11.816296 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb\": container with ID starting with 1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb not found: ID does not exist" containerID="1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.816339 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb"} err="failed to get container status \"1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb\": rpc error: code = NotFound desc = could not find container \"1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb\": container with ID starting with 1f13df56eef4ad7c4ac39dbf9b2b561527443427b7a2d3b3849089a36d9984cb not found: ID does not exist" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.816368 4914 scope.go:117] "RemoveContainer" containerID="92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd" Nov 22 14:52:11 crc kubenswrapper[4914]: E1122 14:52:11.816725 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd\": container with ID starting with 92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd not found: ID does not exist" containerID="92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.816757 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd"} err="failed to get container status \"92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd\": rpc error: code = NotFound desc = could not find container \"92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd\": container with ID starting with 92bff96553d71e9414f1fc60fecb4ebf3e45c9926fc2c633392c32c8fd5812fd not found: ID does not exist" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.816775 4914 scope.go:117] "RemoveContainer" containerID="7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f" Nov 22 14:52:11 crc kubenswrapper[4914]: E1122 14:52:11.817136 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f\": container with ID starting with 7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f not found: ID does not exist" containerID="7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.817165 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f"} err="failed to get container status \"7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f\": rpc error: code = NotFound desc = could not find container \"7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f\": container with ID starting with 7686e076ff9c7fc1a6c6a1c0b8f78fd38aaa34305b6e7c397028fcad706fcd1f not found: ID does not exist" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.902283 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10f05c4b-2473-4efd-b408-e9870dff116f" (UID: "10f05c4b-2473-4efd-b408-e9870dff116f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:52:11 crc kubenswrapper[4914]: I1122 14:52:11.977370 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f05c4b-2473-4efd-b408-e9870dff116f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:52:12 crc kubenswrapper[4914]: I1122 14:52:12.036327 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqfm2"] Nov 22 14:52:12 crc kubenswrapper[4914]: I1122 14:52:12.044654 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wqfm2"] Nov 22 14:52:12 crc kubenswrapper[4914]: I1122 14:52:12.875462 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" path="/var/lib/kubelet/pods/10f05c4b-2473-4efd-b408-e9870dff116f/volumes" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.044962 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-vxptv"] Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.056465 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-vxptv"] Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.479956 4914 scope.go:117] "RemoveContainer" containerID="b5bb04c010728c244b89cd21ab4877c3c2544a4f2b0fc11c6b0ac82deea7e9ba" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.500745 4914 scope.go:117] "RemoveContainer" containerID="6a193eb20cffb18b49340e00f99afd5660b33fa54f1a570f6123fd4192399712" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.530944 4914 scope.go:117] "RemoveContainer" containerID="86133d606683f905138091f990f4c074fa56016a969c3824069e36917953961a" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.557028 4914 scope.go:117] "RemoveContainer" containerID="903b9b43db8a3b106ce4025f53afbc4df3b0638aec7d72e3c68f54fada7a9f9c" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.619225 4914 scope.go:117] "RemoveContainer" containerID="19ee2fc85df36491a764da884c242dab024886b3c1f749c5a359a58e0a002aa4" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.652358 4914 scope.go:117] "RemoveContainer" containerID="11f5f0d72a1007462d2b0f5f66d7f9643eb66229ccf9d41656b1517cc77cd893" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.691922 4914 scope.go:117] "RemoveContainer" containerID="5b4657bfc2b93a1c1030a4f89c4f94304acf5fc4d3b2cf2f3b19bdfdedc7f9c1" Nov 22 14:52:19 crc kubenswrapper[4914]: I1122 14:52:19.868686 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:52:19 crc kubenswrapper[4914]: E1122 14:52:19.870181 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:52:20 crc kubenswrapper[4914]: I1122 14:52:20.879141 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7c25c30-9a98-4091-be20-58ae513514bb" path="/var/lib/kubelet/pods/b7c25c30-9a98-4091-be20-58ae513514bb/volumes" Nov 22 14:52:25 crc kubenswrapper[4914]: I1122 14:52:25.042213 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-px7jm"] Nov 22 14:52:25 crc kubenswrapper[4914]: I1122 14:52:25.051745 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-px7jm"] Nov 22 14:52:26 crc kubenswrapper[4914]: I1122 14:52:26.915556 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b65bb30-a8e0-44b3-8400-9bcd7d99a184" path="/var/lib/kubelet/pods/5b65bb30-a8e0-44b3-8400-9bcd7d99a184/volumes" Nov 22 14:52:33 crc kubenswrapper[4914]: I1122 14:52:33.867401 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:52:33 crc kubenswrapper[4914]: E1122 14:52:33.868209 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.883679 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:52:40 crc kubenswrapper[4914]: E1122 14:52:40.884444 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="extract-utilities" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.884456 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="extract-utilities" Nov 22 14:52:40 crc kubenswrapper[4914]: E1122 14:52:40.884476 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="registry-server" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.884484 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="registry-server" Nov 22 14:52:40 crc kubenswrapper[4914]: E1122 14:52:40.884517 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="extract-content" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.884524 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="extract-content" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.884663 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f05c4b-2473-4efd-b408-e9870dff116f" containerName="registry-server" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.885131 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.889545 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-h4nlp" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.889661 4914 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.889863 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.891957 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 22 14:52:40 crc kubenswrapper[4914]: I1122 14:52:40.895280 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.077492 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.077803 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-config\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.077943 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-scripts\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.078104 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7xlt\" (UniqueName: \"kubernetes.io/projected/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-kube-api-access-d7xlt\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.179788 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-scripts\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.179888 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7xlt\" (UniqueName: \"kubernetes.io/projected/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-kube-api-access-d7xlt\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.179943 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.179971 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-config\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.181331 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-config\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.182143 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-scripts\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.187618 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.199337 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7xlt\" (UniqueName: \"kubernetes.io/projected/c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae-kube-api-access-d7xlt\") pod \"openstackclient\" (UID: \"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae\") " pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.253949 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 22 14:52:41 crc kubenswrapper[4914]: I1122 14:52:41.632382 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 22 14:52:42 crc kubenswrapper[4914]: I1122 14:52:42.050033 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae","Type":"ContainerStarted","Data":"db0c6fc91af7b2397e4d41d01b53a5566d8715724739f0ae3e3ff202661775ce"} Nov 22 14:52:42 crc kubenswrapper[4914]: I1122 14:52:42.050304 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae","Type":"ContainerStarted","Data":"9776ee72089588749560bac765a65b4f96fe525c8b4fe0031ae137805e81648e"} Nov 22 14:52:42 crc kubenswrapper[4914]: I1122 14:52:42.072615 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.072599901 podStartE2EDuration="2.072599901s" podCreationTimestamp="2025-11-22 14:52:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 14:52:42.066700947 +0000 UTC m=+1527.875270608" watchObservedRunningTime="2025-11-22 14:52:42.072599901 +0000 UTC m=+1527.881169562" Nov 22 14:52:44 crc kubenswrapper[4914]: I1122 14:52:44.874370 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:52:44 crc kubenswrapper[4914]: E1122 14:52:44.875019 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:52:59 crc kubenswrapper[4914]: I1122 14:52:59.867860 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:52:59 crc kubenswrapper[4914]: E1122 14:52:59.868681 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:53:10 crc kubenswrapper[4914]: I1122 14:53:10.867691 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:53:10 crc kubenswrapper[4914]: E1122 14:53:10.868620 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:53:19 crc kubenswrapper[4914]: I1122 14:53:19.857499 4914 scope.go:117] "RemoveContainer" containerID="bcea9dad0d1a61b1b0b50d60065b2a16a7bba31f6eea656199b42b259311e8ca" Nov 22 14:53:19 crc kubenswrapper[4914]: I1122 14:53:19.893922 4914 scope.go:117] "RemoveContainer" containerID="519c84eac4cdc9cf5fb89c90373219025c0bdd90711c6f90babcbe745dbfe6b4" Nov 22 14:53:22 crc kubenswrapper[4914]: I1122 14:53:22.867850 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:53:22 crc kubenswrapper[4914]: E1122 14:53:22.869127 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:53:37 crc kubenswrapper[4914]: I1122 14:53:37.868859 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:53:37 crc kubenswrapper[4914]: E1122 14:53:37.869905 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:53:48 crc kubenswrapper[4914]: I1122 14:53:48.868172 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:53:48 crc kubenswrapper[4914]: E1122 14:53:48.869122 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:54:02 crc kubenswrapper[4914]: I1122 14:54:02.868274 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:54:02 crc kubenswrapper[4914]: E1122 14:54:02.869957 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.183565 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gbgx6/must-gather-7xr9r"] Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.185662 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.188281 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gbgx6"/"openshift-service-ca.crt" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.188597 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gbgx6"/"kube-root-ca.crt" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.190475 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gbgx6/must-gather-7xr9r"] Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.340449 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvvhc\" (UniqueName: \"kubernetes.io/projected/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-kube-api-access-kvvhc\") pod \"must-gather-7xr9r\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.340560 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-must-gather-output\") pod \"must-gather-7xr9r\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.441993 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvvhc\" (UniqueName: \"kubernetes.io/projected/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-kube-api-access-kvvhc\") pod \"must-gather-7xr9r\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.442116 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-must-gather-output\") pod \"must-gather-7xr9r\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.442824 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-must-gather-output\") pod \"must-gather-7xr9r\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.461215 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvvhc\" (UniqueName: \"kubernetes.io/projected/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-kube-api-access-kvvhc\") pod \"must-gather-7xr9r\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.514990 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:54:11 crc kubenswrapper[4914]: I1122 14:54:11.939417 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gbgx6/must-gather-7xr9r"] Nov 22 14:54:12 crc kubenswrapper[4914]: I1122 14:54:12.814575 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" event={"ID":"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742","Type":"ContainerStarted","Data":"9c9f989909d04e8889cdcfdb2775716fc59778eca04dc40aa6dfe177a37fd4e3"} Nov 22 14:54:17 crc kubenswrapper[4914]: I1122 14:54:17.853692 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" event={"ID":"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742","Type":"ContainerStarted","Data":"ea495a0fc78283455c5447ab092e351ac10e0397ea390e32f7b22cf3f528d255"} Nov 22 14:54:17 crc kubenswrapper[4914]: I1122 14:54:17.867519 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:54:17 crc kubenswrapper[4914]: E1122 14:54:17.867730 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:54:18 crc kubenswrapper[4914]: I1122 14:54:18.860764 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" event={"ID":"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742","Type":"ContainerStarted","Data":"cd445939a4a26f5bf582919c742b8af02c1038207f3830e14c20a8710c3d9b56"} Nov 22 14:54:18 crc kubenswrapper[4914]: I1122 14:54:18.877525 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" podStartSLOduration=2.194369365 podStartE2EDuration="7.877507016s" podCreationTimestamp="2025-11-22 14:54:11 +0000 UTC" firstStartedPulling="2025-11-22 14:54:11.944851009 +0000 UTC m=+1617.753420680" lastFinishedPulling="2025-11-22 14:54:17.62798867 +0000 UTC m=+1623.436558331" observedRunningTime="2025-11-22 14:54:18.874853961 +0000 UTC m=+1624.683423622" watchObservedRunningTime="2025-11-22 14:54:18.877507016 +0000 UTC m=+1624.686076677" Nov 22 14:54:29 crc kubenswrapper[4914]: I1122 14:54:29.868279 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:54:29 crc kubenswrapper[4914]: E1122 14:54:29.870178 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:54:41 crc kubenswrapper[4914]: I1122 14:54:41.867802 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:54:41 crc kubenswrapper[4914]: E1122 14:54:41.868656 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.231186 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/util/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.413364 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/pull/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.434609 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/util/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.444764 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/pull/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.582423 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/util/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.616783 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/pull/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.663407 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_440bfe09fdefc314a717f31a57d7ebfa5bfc58c8ed9bd7a4e0a57fadb4kscr5_d95bcff7-8ac6-44ee-84b1-83c955faa3af/extract/0.log" Nov 22 14:54:52 crc kubenswrapper[4914]: I1122 14:54:52.860307 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/util/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.050780 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/pull/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.060509 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/pull/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.092221 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/util/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.192709 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/util/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.223711 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/extract/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.259888 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62924b48d7c39bd6b89ba2946b1b573a02aaa01e9968a59c70da1ba1d6qnh2p_8d671252-1412-4afd-8c99-ac3540bb8430/pull/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.352283 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/util/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.532473 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/pull/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.538727 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/pull/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.568816 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/util/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.709318 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/util/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.722408 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/extract/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.750662 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7abe4676e9c7174a0976b528ff13527e30f787694a732dea185c78a27cfwpzl_d0a62272-b520-435c-a59e-783429f5638b/pull/0.log" Nov 22 14:54:53 crc kubenswrapper[4914]: I1122 14:54:53.903304 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/util/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.067297 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/util/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.101627 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/pull/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.101773 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/pull/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.253029 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/util/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.254146 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/pull/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.285140 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9485f81de653cffcbbc17f6e3466e4f47082d898a4066abd5239b49b4dv6zxz_4d27e00b-2931-49a6-9921-27c550ebe6e9/extract/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.412600 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/util/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.598684 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/pull/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.601694 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/util/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.627657 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/pull/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.755464 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/extract/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.767950 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/util/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.802683 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5909r6nh_e046d934-906d-4df1-adcb-f8d67d3d01df/pull/0.log" Nov 22 14:54:54 crc kubenswrapper[4914]: I1122 14:54:54.914187 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/util/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.054514 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/util/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.080675 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/pull/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.085938 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/pull/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.232636 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/util/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.252153 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/extract/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.275116 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/util/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.306537 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ab83952fd3ab379d0f5a55f30eeab7d10e010bcd327dbcc65a9d017a8e9gvgb_c1fb4fba-846d-443d-aa98-665954fd7733/pull/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.420499 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/util/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.422270 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/pull/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.465845 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/pull/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.620093 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/pull/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.640521 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/extract/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.644296 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e7d7e4671d7d05874436e844b5cf6d94e49a6a946d772a1ed2117581bcvpkms_992a8cd6-c0fc-42fd-9668-32260b1e9984/util/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.645407 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-bdf8b57db-6vrpq_84f620d7-7666-4076-a141-b856b3548e6b/kube-rbac-proxy/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.846726 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-5jsh9_2daccf25-7333-4068-9666-765c0e09e4f5/registry-server/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.876027 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-bdf8b57db-6vrpq_84f620d7-7666-4076-a141-b856b3548e6b/manager/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.906030 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f7479bbf4-hbvtt_18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a/kube-rbac-proxy/0.log" Nov 22 14:54:55 crc kubenswrapper[4914]: I1122 14:54:55.978073 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f7479bbf4-hbvtt_18b7bc1b-9168-44ae-9ba7-e354ffe3cb3a/manager/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.022547 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-vcfbn_cd925b6d-ee1d-4af2-a157-492958a1d6ce/registry-server/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.074557 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5b4b54c85c-k5v8g_c72d9635-39e7-46df-9121-4f11c17f7ee5/kube-rbac-proxy/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.159363 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5b4b54c85c-k5v8g_c72d9635-39e7-46df-9121-4f11c17f7ee5/manager/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.237348 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-45zxz_61fd0abd-4102-40bd-af29-8f800b1e4ef6/registry-server/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.299487 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-676b6b6b4-h8ksj_94d945b4-bca0-45e6-8c5f-c2e471e2aad5/kube-rbac-proxy/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.355623 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-676b6b6b4-h8ksj_94d945b4-bca0-45e6-8c5f-c2e471e2aad5/manager/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.458967 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-zs5qg_8ef98c25-d009-447a-a560-94871072156d/registry-server/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.526656 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7888767b5c-zp5lg_10460829-745c-493b-938a-e7dbf0de739e/kube-rbac-proxy/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.552415 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7888767b5c-zp5lg_10460829-745c-493b-938a-e7dbf0de739e/manager/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.634720 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-4rkdv_b1d67ef8-0905-4ec7-8b01-75d765f1f9f3/registry-server/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.700302 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-qdh2z_8b4d9e00-9685-4ca8-8b51-4246e6e65d65/operator/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.731374 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-wn6fr_37ed417d-6cea-4336-9cad-218ec6fa8b12/registry-server/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.816087 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-8497df8d9f-s9nlz_8699b0cd-d651-448d-9750-812ef636a23f/kube-rbac-proxy/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.847780 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-8497df8d9f-s9nlz_8699b0cd-d651-448d-9750-812ef636a23f/manager/0.log" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.867640 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:54:56 crc kubenswrapper[4914]: E1122 14:54:56.867857 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:54:56 crc kubenswrapper[4914]: I1122 14:54:56.966837 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-4xw62_b15488c7-932d-4906-8874-d71f48db7265/registry-server/0.log" Nov 22 14:55:08 crc kubenswrapper[4914]: I1122 14:55:08.868439 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:55:08 crc kubenswrapper[4914]: E1122 14:55:08.869342 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:55:10 crc kubenswrapper[4914]: I1122 14:55:10.244867 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-nk4jg_eaa079dd-220a-4941-90e7-958571a207bd/control-plane-machine-set-operator/0.log" Nov 22 14:55:10 crc kubenswrapper[4914]: I1122 14:55:10.346127 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vb7pw_549dbc44-f8b5-406d-9431-f2a382f44635/kube-rbac-proxy/0.log" Nov 22 14:55:10 crc kubenswrapper[4914]: I1122 14:55:10.446105 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vb7pw_549dbc44-f8b5-406d-9431-f2a382f44635/machine-api-operator/0.log" Nov 22 14:55:23 crc kubenswrapper[4914]: I1122 14:55:23.867895 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:55:23 crc kubenswrapper[4914]: E1122 14:55:23.869723 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:55:24 crc kubenswrapper[4914]: I1122 14:55:24.962018 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-58dgb_5a4d4454-920a-4ec8-a619-13f217a6d027/kube-rbac-proxy/0.log" Nov 22 14:55:24 crc kubenswrapper[4914]: I1122 14:55:24.997080 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-58dgb_5a4d4454-920a-4ec8-a619-13f217a6d027/controller/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.182669 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-frr-files/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.313302 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-metrics/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.318168 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-frr-files/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.337588 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-reloader/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.376413 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-reloader/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.575886 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-metrics/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.577440 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-frr-files/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.599527 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-metrics/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.618370 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-reloader/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.838196 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-frr-files/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.847486 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/controller/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.851841 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-metrics/0.log" Nov 22 14:55:25 crc kubenswrapper[4914]: I1122 14:55:25.870272 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/cp-reloader/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.033208 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/frr-metrics/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.034541 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/kube-rbac-proxy/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.094080 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/kube-rbac-proxy-frr/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.246422 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/reloader/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.377769 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-vg5hv_e6c9a7ae-0f1e-4a8e-83d1-2ec7201579c7/frr-k8s-webhook-server/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.430241 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5cb954cd64-99pbm_7addd0c3-a053-491c-b6a4-3a85cc0263ed/manager/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.432527 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vkmkt_6b1e6db9-848b-48a6-bfc4-053961e4ecb3/frr/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.630308 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5496757488-29bhv_7f9c1310-eb65-4286-8466-e671dbac0ec1/webhook-server/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.640923 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ppqw8_499810cd-69d9-49de-aa7f-a7b20ba8f7b5/kube-rbac-proxy/0.log" Nov 22 14:55:26 crc kubenswrapper[4914]: I1122 14:55:26.833288 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ppqw8_499810cd-69d9-49de-aa7f-a7b20ba8f7b5/speaker/0.log" Nov 22 14:55:38 crc kubenswrapper[4914]: I1122 14:55:38.868078 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:55:38 crc kubenswrapper[4914]: E1122 14:55:38.868751 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.403688 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-9d43-account-create-update-wmdvf_fd247cef-eddf-4ee0-b170-e0b530b2dc48/mariadb-account-create-update/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.517037 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-k98mx_a41c5010-2b3a-4229-bc2a-b1b6a09255e4/mariadb-database-create/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.592370 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-k792j_690d7575-7ec4-4487-a2ef-3ae7037c641d/glance-db-sync/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.668839 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_fb97ee27-5cea-4df8-87ff-72f05c460b73/glance-api/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.780587 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_fb97ee27-5cea-4df8-87ff-72f05c460b73/glance-log/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.784099 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_fb97ee27-5cea-4df8-87ff-72f05c460b73/glance-httpd/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.855501 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_0e0e08d2-dd7c-457e-ae5a-a3b350cd8927/glance-api/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.924090 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_0e0e08d2-dd7c-457e-ae5a-a3b350cd8927/glance-httpd/0.log" Nov 22 14:55:39 crc kubenswrapper[4914]: I1122 14:55:39.945912 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_0e0e08d2-dd7c-457e-ae5a-a3b350cd8927/glance-log/0.log" Nov 22 14:55:40 crc kubenswrapper[4914]: I1122 14:55:40.367621 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-9c89d49d9-jhwn4_1c0e5eea-0d5b-44f0-bc96-b4a6d1ed846f/keystone-api/0.log" Nov 22 14:55:40 crc kubenswrapper[4914]: I1122 14:55:40.509912 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_5a5d3cdb-0a84-422f-8195-59a5eaf47282/mysql-bootstrap/0.log" Nov 22 14:55:40 crc kubenswrapper[4914]: I1122 14:55:40.773370 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_5a5d3cdb-0a84-422f-8195-59a5eaf47282/mysql-bootstrap/0.log" Nov 22 14:55:40 crc kubenswrapper[4914]: I1122 14:55:40.776323 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_5a5d3cdb-0a84-422f-8195-59a5eaf47282/galera/0.log" Nov 22 14:55:40 crc kubenswrapper[4914]: I1122 14:55:40.975034 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_1a2b9eb0-6a18-4786-b176-667b3147c524/mysql-bootstrap/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.143356 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_1a2b9eb0-6a18-4786-b176-667b3147c524/mysql-bootstrap/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.198746 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_1a2b9eb0-6a18-4786-b176-667b3147c524/galera/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.353581 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_17cef1a9-a7c2-481e-82b0-6a6589c9462f/mysql-bootstrap/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.543882 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_17cef1a9-a7c2-481e-82b0-6a6589c9462f/galera/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.568893 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_c67b2123-2f5f-484e-9db5-3b74583d4047/memcached/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.610909 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_17cef1a9-a7c2-481e-82b0-6a6589c9462f/mysql-bootstrap/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.703159 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_c0c28a9a-74a3-4a37-a2a0-81d9cf1d00ae/openstackclient/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.813862 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_713f96a8-d1f3-4fba-bd02-f025ac07a4b1/setup-container/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.945386 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_713f96a8-d1f3-4fba-bd02-f025ac07a4b1/setup-container/0.log" Nov 22 14:55:41 crc kubenswrapper[4914]: I1122 14:55:41.980898 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_713f96a8-d1f3-4fba-bd02-f025ac07a4b1/rabbitmq/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.006012 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-547856594f-9j8pz_f7779691-44d2-40cd-8138-03953251ba7c/proxy-httpd/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.155518 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-547856594f-9j8pz_f7779691-44d2-40cd-8138-03953251ba7c/proxy-server/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.177084 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-zfxrw_e9b69042-3ed9-4df0-86d9-70d4fd48bdc4/swift-ring-rebalance/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.323158 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/account-auditor/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.362548 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/account-replicator/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.387544 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/account-reaper/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.389926 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/account-server/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.506171 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/container-auditor/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.516693 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/container-replicator/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.560711 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/container-updater/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.566862 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/container-server/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.693521 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/object-auditor/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.718836 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/object-expirer/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.729072 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/object-replicator/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.753832 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/object-server/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.871326 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/object-updater/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.879677 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/rsync/0.log" Nov 22 14:55:42 crc kubenswrapper[4914]: I1122 14:55:42.902460 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_cfa8c85b-6b20-4bde-a701-06b713f81c06/swift-recon-cron/0.log" Nov 22 14:55:52 crc kubenswrapper[4914]: I1122 14:55:52.868630 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:55:52 crc kubenswrapper[4914]: E1122 14:55:52.869486 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:55:54 crc kubenswrapper[4914]: I1122 14:55:54.669308 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/extract-utilities/0.log" Nov 22 14:55:54 crc kubenswrapper[4914]: I1122 14:55:54.858238 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/extract-content/0.log" Nov 22 14:55:54 crc kubenswrapper[4914]: I1122 14:55:54.866679 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/extract-utilities/0.log" Nov 22 14:55:54 crc kubenswrapper[4914]: I1122 14:55:54.887942 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/extract-content/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.088817 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/extract-content/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.122823 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/extract-utilities/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.271357 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/extract-utilities/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.444988 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lxdjp_a26ac6be-fa89-4b85-89b9-e24dc1065498/registry-server/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.452925 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/extract-utilities/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.523342 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/extract-content/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.527240 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/extract-content/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.666453 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/extract-utilities/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.680018 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/extract-content/0.log" Nov 22 14:55:55 crc kubenswrapper[4914]: I1122 14:55:55.868136 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/util/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.098537 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/util/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.138459 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/pull/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.145533 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-xvnv6_73e533fa-1996-4a43-949d-f5a1ea000dc7/registry-server/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.173442 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/pull/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.320440 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/util/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.323918 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/extract/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.326266 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6wh8nr_a165a22d-31b5-46dc-835a-7e735a35da62/pull/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.498922 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hvrzj_096f3cee-1a5e-4891-8b45-f1c00531807b/marketplace-operator/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.509304 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/extract-utilities/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.633899 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/extract-utilities/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.652629 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/extract-content/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.671614 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/extract-content/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.810318 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/extract-utilities/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.825196 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/extract-content/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.893942 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pgm5g_24938bf2-538b-4d51-9ac5-db5947989fa2/registry-server/0.log" Nov 22 14:55:56 crc kubenswrapper[4914]: I1122 14:55:56.967373 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/extract-utilities/0.log" Nov 22 14:55:57 crc kubenswrapper[4914]: I1122 14:55:57.143106 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/extract-content/0.log" Nov 22 14:55:57 crc kubenswrapper[4914]: I1122 14:55:57.146649 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/extract-content/0.log" Nov 22 14:55:57 crc kubenswrapper[4914]: I1122 14:55:57.149681 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/extract-utilities/0.log" Nov 22 14:55:57 crc kubenswrapper[4914]: I1122 14:55:57.333263 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/extract-utilities/0.log" Nov 22 14:55:57 crc kubenswrapper[4914]: I1122 14:55:57.370021 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/extract-content/0.log" Nov 22 14:55:57 crc kubenswrapper[4914]: I1122 14:55:57.638145 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jh4qr_94fd4ac6-3c90-45c1-9d44-4f8d6977be8b/registry-server/0.log" Nov 22 14:56:04 crc kubenswrapper[4914]: I1122 14:56:04.877515 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:56:04 crc kubenswrapper[4914]: E1122 14:56:04.878501 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:56:12 crc kubenswrapper[4914]: I1122 14:56:12.060796 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-k98mx"] Nov 22 14:56:12 crc kubenswrapper[4914]: I1122 14:56:12.067015 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-9d43-account-create-update-wmdvf"] Nov 22 14:56:12 crc kubenswrapper[4914]: I1122 14:56:12.072728 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-k98mx"] Nov 22 14:56:12 crc kubenswrapper[4914]: I1122 14:56:12.078609 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-9d43-account-create-update-wmdvf"] Nov 22 14:56:12 crc kubenswrapper[4914]: I1122 14:56:12.876066 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a41c5010-2b3a-4229-bc2a-b1b6a09255e4" path="/var/lib/kubelet/pods/a41c5010-2b3a-4229-bc2a-b1b6a09255e4/volumes" Nov 22 14:56:12 crc kubenswrapper[4914]: I1122 14:56:12.877044 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd247cef-eddf-4ee0-b170-e0b530b2dc48" path="/var/lib/kubelet/pods/fd247cef-eddf-4ee0-b170-e0b530b2dc48/volumes" Nov 22 14:56:17 crc kubenswrapper[4914]: I1122 14:56:17.867947 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:56:17 crc kubenswrapper[4914]: E1122 14:56:17.868722 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:56:19 crc kubenswrapper[4914]: I1122 14:56:19.975033 4914 scope.go:117] "RemoveContainer" containerID="222158e92ce36c73b700992714d8f368e86766d28c2539a4a7e1886b1ecd71f4" Nov 22 14:56:20 crc kubenswrapper[4914]: I1122 14:56:20.012262 4914 scope.go:117] "RemoveContainer" containerID="bcc3d47142545eb7dccb1d5341025fe89b2e6772130d3f6a8d7cfed8226b0946" Nov 22 14:56:20 crc kubenswrapper[4914]: I1122 14:56:20.045781 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-k792j"] Nov 22 14:56:20 crc kubenswrapper[4914]: I1122 14:56:20.052528 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-k792j"] Nov 22 14:56:20 crc kubenswrapper[4914]: I1122 14:56:20.884750 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690d7575-7ec4-4487-a2ef-3ae7037c641d" path="/var/lib/kubelet/pods/690d7575-7ec4-4487-a2ef-3ae7037c641d/volumes" Nov 22 14:56:29 crc kubenswrapper[4914]: I1122 14:56:29.869094 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:56:29 crc kubenswrapper[4914]: E1122 14:56:29.870160 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:56:42 crc kubenswrapper[4914]: I1122 14:56:42.872063 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:56:42 crc kubenswrapper[4914]: E1122 14:56:42.872783 4914 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7khd_openshift-machine-config-operator(b8b64a09-b0a3-4776-93f7-01c4ea3e1388)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" Nov 22 14:56:56 crc kubenswrapper[4914]: I1122 14:56:56.869792 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 14:56:57 crc kubenswrapper[4914]: I1122 14:56:57.319895 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"c94ee23de6e4b6081778414394aa5358d18916bf36e0084a5c16ca5fba83c464"} Nov 22 14:56:59 crc kubenswrapper[4914]: I1122 14:56:59.334833 4914 generic.go:334] "Generic (PLEG): container finished" podID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerID="ea495a0fc78283455c5447ab092e351ac10e0397ea390e32f7b22cf3f528d255" exitCode=0 Nov 22 14:56:59 crc kubenswrapper[4914]: I1122 14:56:59.335012 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" event={"ID":"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742","Type":"ContainerDied","Data":"ea495a0fc78283455c5447ab092e351ac10e0397ea390e32f7b22cf3f528d255"} Nov 22 14:56:59 crc kubenswrapper[4914]: I1122 14:56:59.335907 4914 scope.go:117] "RemoveContainer" containerID="ea495a0fc78283455c5447ab092e351ac10e0397ea390e32f7b22cf3f528d255" Nov 22 14:57:00 crc kubenswrapper[4914]: I1122 14:57:00.064274 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gbgx6_must-gather-7xr9r_e0c8b5a9-b9bb-4cb4-bfd4-72192daba742/gather/0.log" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.086237 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gbgx6/must-gather-7xr9r"] Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.086963 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="copy" containerID="cri-o://cd445939a4a26f5bf582919c742b8af02c1038207f3830e14c20a8710c3d9b56" gracePeriod=2 Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.106808 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gbgx6/must-gather-7xr9r"] Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.399372 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gbgx6_must-gather-7xr9r_e0c8b5a9-b9bb-4cb4-bfd4-72192daba742/copy/0.log" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.400049 4914 generic.go:334] "Generic (PLEG): container finished" podID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerID="cd445939a4a26f5bf582919c742b8af02c1038207f3830e14c20a8710c3d9b56" exitCode=143 Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.400096 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c9f989909d04e8889cdcfdb2775716fc59778eca04dc40aa6dfe177a37fd4e3" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.440258 4914 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gbgx6_must-gather-7xr9r_e0c8b5a9-b9bb-4cb4-bfd4-72192daba742/copy/0.log" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.440662 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.496269 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-must-gather-output\") pod \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.496327 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvvhc\" (UniqueName: \"kubernetes.io/projected/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-kube-api-access-kvvhc\") pod \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\" (UID: \"e0c8b5a9-b9bb-4cb4-bfd4-72192daba742\") " Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.509010 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-kube-api-access-kvvhc" (OuterVolumeSpecName: "kube-api-access-kvvhc") pod "e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" (UID: "e0c8b5a9-b9bb-4cb4-bfd4-72192daba742"). InnerVolumeSpecName "kube-api-access-kvvhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.566383 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" (UID: "e0c8b5a9-b9bb-4cb4-bfd4-72192daba742"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.597659 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvvhc\" (UniqueName: \"kubernetes.io/projected/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-kube-api-access-kvvhc\") on node \"crc\" DevicePath \"\"" Nov 22 14:57:07 crc kubenswrapper[4914]: I1122 14:57:07.597689 4914 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 22 14:57:08 crc kubenswrapper[4914]: I1122 14:57:08.405846 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gbgx6/must-gather-7xr9r" Nov 22 14:57:08 crc kubenswrapper[4914]: I1122 14:57:08.875943 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" path="/var/lib/kubelet/pods/e0c8b5a9-b9bb-4cb4-bfd4-72192daba742/volumes" Nov 22 14:57:20 crc kubenswrapper[4914]: I1122 14:57:20.160775 4914 scope.go:117] "RemoveContainer" containerID="d038c3aca1f7b73e826709fef7de00239b42f3f3ab5c1849134c8d601e03d7d1" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.594825 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jpcsv"] Nov 22 14:58:27 crc kubenswrapper[4914]: E1122 14:58:27.596064 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="gather" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.596081 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="gather" Nov 22 14:58:27 crc kubenswrapper[4914]: E1122 14:58:27.596109 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="copy" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.596118 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="copy" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.596304 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="gather" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.596319 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c8b5a9-b9bb-4cb4-bfd4-72192daba742" containerName="copy" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.597531 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.621484 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpcsv"] Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.792319 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2w52\" (UniqueName: \"kubernetes.io/projected/5ff06e20-b8d8-41cf-973d-50e5513ebda5-kube-api-access-r2w52\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.792657 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-catalog-content\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.792701 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-utilities\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.894030 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2w52\" (UniqueName: \"kubernetes.io/projected/5ff06e20-b8d8-41cf-973d-50e5513ebda5-kube-api-access-r2w52\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.894161 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-catalog-content\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.894243 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-utilities\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.894769 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-catalog-content\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.894787 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-utilities\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.914520 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2w52\" (UniqueName: \"kubernetes.io/projected/5ff06e20-b8d8-41cf-973d-50e5513ebda5-kube-api-access-r2w52\") pod \"redhat-marketplace-jpcsv\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:27 crc kubenswrapper[4914]: I1122 14:58:27.940492 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:28 crc kubenswrapper[4914]: I1122 14:58:28.361685 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpcsv"] Nov 22 14:58:29 crc kubenswrapper[4914]: I1122 14:58:29.068307 4914 generic.go:334] "Generic (PLEG): container finished" podID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerID="0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f" exitCode=0 Nov 22 14:58:29 crc kubenswrapper[4914]: I1122 14:58:29.068382 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpcsv" event={"ID":"5ff06e20-b8d8-41cf-973d-50e5513ebda5","Type":"ContainerDied","Data":"0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f"} Nov 22 14:58:29 crc kubenswrapper[4914]: I1122 14:58:29.068431 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpcsv" event={"ID":"5ff06e20-b8d8-41cf-973d-50e5513ebda5","Type":"ContainerStarted","Data":"ee11f31c5529ad346d0636915f16b72d7729898f0416ebdfb88d7847ba7ecf24"} Nov 22 14:58:29 crc kubenswrapper[4914]: I1122 14:58:29.072100 4914 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 22 14:58:30 crc kubenswrapper[4914]: I1122 14:58:30.078344 4914 generic.go:334] "Generic (PLEG): container finished" podID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerID="8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3" exitCode=0 Nov 22 14:58:30 crc kubenswrapper[4914]: I1122 14:58:30.078447 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpcsv" event={"ID":"5ff06e20-b8d8-41cf-973d-50e5513ebda5","Type":"ContainerDied","Data":"8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3"} Nov 22 14:58:31 crc kubenswrapper[4914]: I1122 14:58:31.098973 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpcsv" event={"ID":"5ff06e20-b8d8-41cf-973d-50e5513ebda5","Type":"ContainerStarted","Data":"8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721"} Nov 22 14:58:31 crc kubenswrapper[4914]: I1122 14:58:31.130331 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jpcsv" podStartSLOduration=2.74519458 podStartE2EDuration="4.130297425s" podCreationTimestamp="2025-11-22 14:58:27 +0000 UTC" firstStartedPulling="2025-11-22 14:58:29.071345502 +0000 UTC m=+1874.879915173" lastFinishedPulling="2025-11-22 14:58:30.456448357 +0000 UTC m=+1876.265018018" observedRunningTime="2025-11-22 14:58:31.129477385 +0000 UTC m=+1876.938047056" watchObservedRunningTime="2025-11-22 14:58:31.130297425 +0000 UTC m=+1876.938867106" Nov 22 14:58:37 crc kubenswrapper[4914]: I1122 14:58:37.941570 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:37 crc kubenswrapper[4914]: I1122 14:58:37.942355 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:38 crc kubenswrapper[4914]: I1122 14:58:38.004147 4914 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:38 crc kubenswrapper[4914]: I1122 14:58:38.187811 4914 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:38 crc kubenswrapper[4914]: I1122 14:58:38.236258 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpcsv"] Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.165983 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jpcsv" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="registry-server" containerID="cri-o://8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721" gracePeriod=2 Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.596083 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.797429 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-utilities\") pod \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.797509 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-catalog-content\") pod \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.797579 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2w52\" (UniqueName: \"kubernetes.io/projected/5ff06e20-b8d8-41cf-973d-50e5513ebda5-kube-api-access-r2w52\") pod \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\" (UID: \"5ff06e20-b8d8-41cf-973d-50e5513ebda5\") " Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.798345 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-utilities" (OuterVolumeSpecName: "utilities") pod "5ff06e20-b8d8-41cf-973d-50e5513ebda5" (UID: "5ff06e20-b8d8-41cf-973d-50e5513ebda5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.803406 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff06e20-b8d8-41cf-973d-50e5513ebda5-kube-api-access-r2w52" (OuterVolumeSpecName: "kube-api-access-r2w52") pod "5ff06e20-b8d8-41cf-973d-50e5513ebda5" (UID: "5ff06e20-b8d8-41cf-973d-50e5513ebda5"). InnerVolumeSpecName "kube-api-access-r2w52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.814050 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ff06e20-b8d8-41cf-973d-50e5513ebda5" (UID: "5ff06e20-b8d8-41cf-973d-50e5513ebda5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.899406 4914 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-utilities\") on node \"crc\" DevicePath \"\"" Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.899450 4914 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff06e20-b8d8-41cf-973d-50e5513ebda5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 22 14:58:40 crc kubenswrapper[4914]: I1122 14:58:40.899463 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2w52\" (UniqueName: \"kubernetes.io/projected/5ff06e20-b8d8-41cf-973d-50e5513ebda5-kube-api-access-r2w52\") on node \"crc\" DevicePath \"\"" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.174271 4914 generic.go:334] "Generic (PLEG): container finished" podID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerID="8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721" exitCode=0 Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.174317 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpcsv" event={"ID":"5ff06e20-b8d8-41cf-973d-50e5513ebda5","Type":"ContainerDied","Data":"8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721"} Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.174334 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpcsv" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.174349 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpcsv" event={"ID":"5ff06e20-b8d8-41cf-973d-50e5513ebda5","Type":"ContainerDied","Data":"ee11f31c5529ad346d0636915f16b72d7729898f0416ebdfb88d7847ba7ecf24"} Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.174369 4914 scope.go:117] "RemoveContainer" containerID="8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.194467 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpcsv"] Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.201560 4914 scope.go:117] "RemoveContainer" containerID="8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.201633 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpcsv"] Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.222204 4914 scope.go:117] "RemoveContainer" containerID="0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.248855 4914 scope.go:117] "RemoveContainer" containerID="8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721" Nov 22 14:58:41 crc kubenswrapper[4914]: E1122 14:58:41.249467 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721\": container with ID starting with 8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721 not found: ID does not exist" containerID="8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.249516 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721"} err="failed to get container status \"8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721\": rpc error: code = NotFound desc = could not find container \"8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721\": container with ID starting with 8f368ee5d688803dca1d28f0d116f769ebcb19165321513acae161974e393721 not found: ID does not exist" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.249543 4914 scope.go:117] "RemoveContainer" containerID="8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3" Nov 22 14:58:41 crc kubenswrapper[4914]: E1122 14:58:41.250130 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3\": container with ID starting with 8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3 not found: ID does not exist" containerID="8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.250173 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3"} err="failed to get container status \"8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3\": rpc error: code = NotFound desc = could not find container \"8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3\": container with ID starting with 8ab0f62264480bf473586f810dc15360f08b06a9f1ee483d4041e941168d2aa3 not found: ID does not exist" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.250208 4914 scope.go:117] "RemoveContainer" containerID="0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f" Nov 22 14:58:41 crc kubenswrapper[4914]: E1122 14:58:41.251384 4914 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f\": container with ID starting with 0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f not found: ID does not exist" containerID="0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f" Nov 22 14:58:41 crc kubenswrapper[4914]: I1122 14:58:41.251414 4914 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f"} err="failed to get container status \"0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f\": rpc error: code = NotFound desc = could not find container \"0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f\": container with ID starting with 0bf8b93a991d5f23261d8363bcf00397add390ca287aa929c18690235c1eaf9f not found: ID does not exist" Nov 22 14:58:42 crc kubenswrapper[4914]: I1122 14:58:42.882377 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" path="/var/lib/kubelet/pods/5ff06e20-b8d8-41cf-973d-50e5513ebda5/volumes" Nov 22 14:59:22 crc kubenswrapper[4914]: I1122 14:59:22.042515 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:59:22 crc kubenswrapper[4914]: I1122 14:59:22.043096 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 14:59:52 crc kubenswrapper[4914]: I1122 14:59:52.042338 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 14:59:52 crc kubenswrapper[4914]: I1122 14:59:52.042937 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.147931 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56"] Nov 22 15:00:00 crc kubenswrapper[4914]: E1122 15:00:00.158379 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="extract-content" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.158641 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="extract-content" Nov 22 15:00:00 crc kubenswrapper[4914]: E1122 15:00:00.158710 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="registry-server" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.158779 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="registry-server" Nov 22 15:00:00 crc kubenswrapper[4914]: E1122 15:00:00.158849 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="extract-utilities" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.158937 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="extract-utilities" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.159135 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff06e20-b8d8-41cf-973d-50e5513ebda5" containerName="registry-server" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.159599 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb"] Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.159682 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.161985 4914 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.162851 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.163085 4914 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.164129 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9"] Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.164854 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.170607 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56"] Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.178438 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9"] Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.191160 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb"] Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.317909 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38f23993-276e-4d56-ba86-218aab443898-config-volume\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318234 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318260 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vskjs\" (UniqueName: \"kubernetes.io/projected/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-kube-api-access-vskjs\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318297 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24hs\" (UniqueName: \"kubernetes.io/projected/38f23993-276e-4d56-ba86-218aab443898-kube-api-access-m24hs\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318326 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/7259e62c-d061-4d29-a3c3-c162bb8af311-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318345 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26z59\" (UniqueName: \"kubernetes.io/projected/7259e62c-d061-4d29-a3c3-c162bb8af311-kube-api-access-26z59\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318371 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318410 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.318441 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38f23993-276e-4d56-ba86-218aab443898-secret-volume\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.338351 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.346596 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.419776 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24hs\" (UniqueName: \"kubernetes.io/projected/38f23993-276e-4d56-ba86-218aab443898-kube-api-access-m24hs\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.419829 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/7259e62c-d061-4d29-a3c3-c162bb8af311-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.419889 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26z59\" (UniqueName: \"kubernetes.io/projected/7259e62c-d061-4d29-a3c3-c162bb8af311-kube-api-access-26z59\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.419923 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.419963 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38f23993-276e-4d56-ba86-218aab443898-secret-volume\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.420013 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38f23993-276e-4d56-ba86-218aab443898-config-volume\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.420035 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vskjs\" (UniqueName: \"kubernetes.io/projected/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-kube-api-access-vskjs\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.425293 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/7259e62c-d061-4d29-a3c3-c162bb8af311-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.432860 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38f23993-276e-4d56-ba86-218aab443898-config-volume\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.435431 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.435620 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38f23993-276e-4d56-ba86-218aab443898-secret-volume\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.438793 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24hs\" (UniqueName: \"kubernetes.io/projected/38f23993-276e-4d56-ba86-218aab443898-kube-api-access-m24hs\") pod \"collect-profiles-29397060-dsg56\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.441356 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26z59\" (UniqueName: \"kubernetes.io/projected/7259e62c-d061-4d29-a3c3-c162bb8af311-kube-api-access-26z59\") pod \"glance-cache-glance-default-external-api-0-cleaner-29397062ddt9\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.443440 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vskjs\" (UniqueName: \"kubernetes.io/projected/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-kube-api-access-vskjs\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2939706288pb\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.497106 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.505868 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.521560 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:00 crc kubenswrapper[4914]: I1122 15:00:00.929688 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56"] Nov 22 15:00:01 crc kubenswrapper[4914]: W1122 15:00:01.000000 4914 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7259e62c_d061_4d29_a3c3_c162bb8af311.slice/crio-1345c2df24e16c79d21d240b5aef67e29b4c0c851a5aabd1cce5ddc4daad8d5f WatchSource:0}: Error finding container 1345c2df24e16c79d21d240b5aef67e29b4c0c851a5aabd1cce5ddc4daad8d5f: Status 404 returned error can't find the container with id 1345c2df24e16c79d21d240b5aef67e29b4c0c851a5aabd1cce5ddc4daad8d5f Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.003557 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9"] Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.016109 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb"] Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.787727 4914 generic.go:334] "Generic (PLEG): container finished" podID="38f23993-276e-4d56-ba86-218aab443898" containerID="130f2fa9b47ce22ee202f1771a5aabc53c544781bf9b5437f5a600f2e8bc7f3e" exitCode=0 Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.787905 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" event={"ID":"38f23993-276e-4d56-ba86-218aab443898","Type":"ContainerDied","Data":"130f2fa9b47ce22ee202f1771a5aabc53c544781bf9b5437f5a600f2e8bc7f3e"} Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.787974 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" event={"ID":"38f23993-276e-4d56-ba86-218aab443898","Type":"ContainerStarted","Data":"4083cda074921ed1de85aec2e1bcb4238b015afe8c86bf2148b74ed0f5c66072"} Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.790109 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" event={"ID":"8b806fa9-949d-44d4-8eb7-22a2e7f97a92","Type":"ContainerStarted","Data":"04d8c246c001e827da2475adb7caf63cf884f46355986d04ab449fbb5a063e48"} Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.790319 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" event={"ID":"8b806fa9-949d-44d4-8eb7-22a2e7f97a92","Type":"ContainerStarted","Data":"d199e182aea22bcf599809bf192fe2f677cfb608eb28287257f2f5531a454fdb"} Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.792143 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" event={"ID":"7259e62c-d061-4d29-a3c3-c162bb8af311","Type":"ContainerStarted","Data":"9e13a8a7f7d27dcc672fcb625028a1bbd1cecd0e60bfcc8e0f9ead191d6a0c0d"} Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.792171 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" event={"ID":"7259e62c-d061-4d29-a3c3-c162bb8af311","Type":"ContainerStarted","Data":"1345c2df24e16c79d21d240b5aef67e29b4c0c851a5aabd1cce5ddc4daad8d5f"} Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.820918 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" podStartSLOduration=1.8208993169999999 podStartE2EDuration="1.820899317s" podCreationTimestamp="2025-11-22 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 15:00:01.814856348 +0000 UTC m=+1967.623426009" watchObservedRunningTime="2025-11-22 15:00:01.820899317 +0000 UTC m=+1967.629468978" Nov 22 15:00:01 crc kubenswrapper[4914]: I1122 15:00:01.829445 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" podStartSLOduration=1.829428956 podStartE2EDuration="1.829428956s" podCreationTimestamp="2025-11-22 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 15:00:01.828712868 +0000 UTC m=+1967.637282569" watchObservedRunningTime="2025-11-22 15:00:01.829428956 +0000 UTC m=+1967.637998617" Nov 22 15:00:02 crc kubenswrapper[4914]: I1122 15:00:02.803330 4914 generic.go:334] "Generic (PLEG): container finished" podID="8b806fa9-949d-44d4-8eb7-22a2e7f97a92" containerID="04d8c246c001e827da2475adb7caf63cf884f46355986d04ab449fbb5a063e48" exitCode=0 Nov 22 15:00:02 crc kubenswrapper[4914]: I1122 15:00:02.803407 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" event={"ID":"8b806fa9-949d-44d4-8eb7-22a2e7f97a92","Type":"ContainerDied","Data":"04d8c246c001e827da2475adb7caf63cf884f46355986d04ab449fbb5a063e48"} Nov 22 15:00:02 crc kubenswrapper[4914]: I1122 15:00:02.804983 4914 generic.go:334] "Generic (PLEG): container finished" podID="7259e62c-d061-4d29-a3c3-c162bb8af311" containerID="9e13a8a7f7d27dcc672fcb625028a1bbd1cecd0e60bfcc8e0f9ead191d6a0c0d" exitCode=0 Nov 22 15:00:02 crc kubenswrapper[4914]: I1122 15:00:02.805370 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" event={"ID":"7259e62c-d061-4d29-a3c3-c162bb8af311","Type":"ContainerDied","Data":"9e13a8a7f7d27dcc672fcb625028a1bbd1cecd0e60bfcc8e0f9ead191d6a0c0d"} Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.100824 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.268936 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m24hs\" (UniqueName: \"kubernetes.io/projected/38f23993-276e-4d56-ba86-218aab443898-kube-api-access-m24hs\") pod \"38f23993-276e-4d56-ba86-218aab443898\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.269050 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38f23993-276e-4d56-ba86-218aab443898-secret-volume\") pod \"38f23993-276e-4d56-ba86-218aab443898\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.269106 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38f23993-276e-4d56-ba86-218aab443898-config-volume\") pod \"38f23993-276e-4d56-ba86-218aab443898\" (UID: \"38f23993-276e-4d56-ba86-218aab443898\") " Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.269781 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38f23993-276e-4d56-ba86-218aab443898-config-volume" (OuterVolumeSpecName: "config-volume") pod "38f23993-276e-4d56-ba86-218aab443898" (UID: "38f23993-276e-4d56-ba86-218aab443898"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.274927 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38f23993-276e-4d56-ba86-218aab443898-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "38f23993-276e-4d56-ba86-218aab443898" (UID: "38f23993-276e-4d56-ba86-218aab443898"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.281487 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38f23993-276e-4d56-ba86-218aab443898-kube-api-access-m24hs" (OuterVolumeSpecName: "kube-api-access-m24hs") pod "38f23993-276e-4d56-ba86-218aab443898" (UID: "38f23993-276e-4d56-ba86-218aab443898"). InnerVolumeSpecName "kube-api-access-m24hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.370839 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m24hs\" (UniqueName: \"kubernetes.io/projected/38f23993-276e-4d56-ba86-218aab443898-kube-api-access-m24hs\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.370887 4914 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38f23993-276e-4d56-ba86-218aab443898-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.370896 4914 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38f23993-276e-4d56-ba86-218aab443898-config-volume\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.813798 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" event={"ID":"38f23993-276e-4d56-ba86-218aab443898","Type":"ContainerDied","Data":"4083cda074921ed1de85aec2e1bcb4238b015afe8c86bf2148b74ed0f5c66072"} Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.813911 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4083cda074921ed1de85aec2e1bcb4238b015afe8c86bf2148b74ed0f5c66072" Nov 22 15:00:03 crc kubenswrapper[4914]: I1122 15:00:03.813984 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29397060-dsg56" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.158627 4914 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8"] Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.164208 4914 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29397015-9smj8"] Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.164730 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.170214 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.284588 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.284663 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vskjs\" (UniqueName: \"kubernetes.io/projected/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-kube-api-access-vskjs\") pod \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.284689 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/7259e62c-d061-4d29-a3c3-c162bb8af311-image-cache-config-data\") pod \"7259e62c-d061-4d29-a3c3-c162bb8af311\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.284706 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"7259e62c-d061-4d29-a3c3-c162bb8af311\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.284761 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-image-cache-config-data\") pod \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\" (UID: \"8b806fa9-949d-44d4-8eb7-22a2e7f97a92\") " Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.284787 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26z59\" (UniqueName: \"kubernetes.io/projected/7259e62c-d061-4d29-a3c3-c162bb8af311-kube-api-access-26z59\") pod \"7259e62c-d061-4d29-a3c3-c162bb8af311\" (UID: \"7259e62c-d061-4d29-a3c3-c162bb8af311\") " Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.287902 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "8b806fa9-949d-44d4-8eb7-22a2e7f97a92" (UID: "8b806fa9-949d-44d4-8eb7-22a2e7f97a92"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.287937 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-kube-api-access-vskjs" (OuterVolumeSpecName: "kube-api-access-vskjs") pod "8b806fa9-949d-44d4-8eb7-22a2e7f97a92" (UID: "8b806fa9-949d-44d4-8eb7-22a2e7f97a92"). InnerVolumeSpecName "kube-api-access-vskjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.288324 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7259e62c-d061-4d29-a3c3-c162bb8af311-kube-api-access-26z59" (OuterVolumeSpecName: "kube-api-access-26z59") pod "7259e62c-d061-4d29-a3c3-c162bb8af311" (UID: "7259e62c-d061-4d29-a3c3-c162bb8af311"). InnerVolumeSpecName "kube-api-access-26z59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.288909 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "7259e62c-d061-4d29-a3c3-c162bb8af311" (UID: "7259e62c-d061-4d29-a3c3-c162bb8af311"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.289433 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "8b806fa9-949d-44d4-8eb7-22a2e7f97a92" (UID: "8b806fa9-949d-44d4-8eb7-22a2e7f97a92"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.290661 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7259e62c-d061-4d29-a3c3-c162bb8af311-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "7259e62c-d061-4d29-a3c3-c162bb8af311" (UID: "7259e62c-d061-4d29-a3c3-c162bb8af311"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.387195 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vskjs\" (UniqueName: \"kubernetes.io/projected/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-kube-api-access-vskjs\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.387275 4914 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/7259e62c-d061-4d29-a3c3-c162bb8af311-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.387290 4914 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/8b806fa9-949d-44d4-8eb7-22a2e7f97a92-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.387301 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26z59\" (UniqueName: \"kubernetes.io/projected/7259e62c-d061-4d29-a3c3-c162bb8af311-kube-api-access-26z59\") on node \"crc\" DevicePath \"\"" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.838102 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" event={"ID":"8b806fa9-949d-44d4-8eb7-22a2e7f97a92","Type":"ContainerDied","Data":"d199e182aea22bcf599809bf192fe2f677cfb608eb28287257f2f5531a454fdb"} Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.838167 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d199e182aea22bcf599809bf192fe2f677cfb608eb28287257f2f5531a454fdb" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.838132 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2939706288pb" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.843977 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" event={"ID":"7259e62c-d061-4d29-a3c3-c162bb8af311","Type":"ContainerDied","Data":"1345c2df24e16c79d21d240b5aef67e29b4c0c851a5aabd1cce5ddc4daad8d5f"} Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.844025 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1345c2df24e16c79d21d240b5aef67e29b4c0c851a5aabd1cce5ddc4daad8d5f" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.844122 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-29397062ddt9" Nov 22 15:00:04 crc kubenswrapper[4914]: I1122 15:00:04.877418 4914 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e44ee621-6ddb-4ee6-bf5a-07215d61a280" path="/var/lib/kubelet/pods/e44ee621-6ddb-4ee6-bf5a-07215d61a280/volumes" Nov 22 15:00:20 crc kubenswrapper[4914]: I1122 15:00:20.317509 4914 scope.go:117] "RemoveContainer" containerID="ea495a0fc78283455c5447ab092e351ac10e0397ea390e32f7b22cf3f528d255" Nov 22 15:00:20 crc kubenswrapper[4914]: I1122 15:00:20.373212 4914 scope.go:117] "RemoveContainer" containerID="cd445939a4a26f5bf582919c742b8af02c1038207f3830e14c20a8710c3d9b56" Nov 22 15:00:20 crc kubenswrapper[4914]: I1122 15:00:20.396799 4914 scope.go:117] "RemoveContainer" containerID="e5bfe0155e6c48f5dae83c44bf076f529cfa1827f58e5b7f2300e3a0de019c2b" Nov 22 15:00:22 crc kubenswrapper[4914]: I1122 15:00:22.042237 4914 patch_prober.go:28] interesting pod/machine-config-daemon-g7khd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 22 15:00:22 crc kubenswrapper[4914]: I1122 15:00:22.042583 4914 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 22 15:00:22 crc kubenswrapper[4914]: I1122 15:00:22.042630 4914 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" Nov 22 15:00:22 crc kubenswrapper[4914]: I1122 15:00:22.043357 4914 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c94ee23de6e4b6081778414394aa5358d18916bf36e0084a5c16ca5fba83c464"} pod="openshift-machine-config-operator/machine-config-daemon-g7khd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 22 15:00:22 crc kubenswrapper[4914]: I1122 15:00:22.043424 4914 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" podUID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerName="machine-config-daemon" containerID="cri-o://c94ee23de6e4b6081778414394aa5358d18916bf36e0084a5c16ca5fba83c464" gracePeriod=600 Nov 22 15:00:23 crc kubenswrapper[4914]: I1122 15:00:23.021498 4914 generic.go:334] "Generic (PLEG): container finished" podID="b8b64a09-b0a3-4776-93f7-01c4ea3e1388" containerID="c94ee23de6e4b6081778414394aa5358d18916bf36e0084a5c16ca5fba83c464" exitCode=0 Nov 22 15:00:23 crc kubenswrapper[4914]: I1122 15:00:23.021611 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerDied","Data":"c94ee23de6e4b6081778414394aa5358d18916bf36e0084a5c16ca5fba83c464"} Nov 22 15:00:23 crc kubenswrapper[4914]: I1122 15:00:23.021690 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7khd" event={"ID":"b8b64a09-b0a3-4776-93f7-01c4ea3e1388","Type":"ContainerStarted","Data":"3955855e0ccb67cc33e87114e557d04e280a24466229b049981af603d83d84bc"} Nov 22 15:00:23 crc kubenswrapper[4914]: I1122 15:00:23.021722 4914 scope.go:117] "RemoveContainer" containerID="fe87350e8f5dbefa873c9b6165cc325d4c84842c76ba46a72bfb83f74a2753de" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.164078 4914 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cron-29397061-p7j6w"] Nov 22 15:01:00 crc kubenswrapper[4914]: E1122 15:01:00.165068 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38f23993-276e-4d56-ba86-218aab443898" containerName="collect-profiles" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.165089 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="38f23993-276e-4d56-ba86-218aab443898" containerName="collect-profiles" Nov 22 15:01:00 crc kubenswrapper[4914]: E1122 15:01:00.165132 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b806fa9-949d-44d4-8eb7-22a2e7f97a92" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.165146 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b806fa9-949d-44d4-8eb7-22a2e7f97a92" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 22 15:01:00 crc kubenswrapper[4914]: E1122 15:01:00.165177 4914 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7259e62c-d061-4d29-a3c3-c162bb8af311" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.165190 4914 state_mem.go:107] "Deleted CPUSet assignment" podUID="7259e62c-d061-4d29-a3c3-c162bb8af311" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.165420 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="38f23993-276e-4d56-ba86-218aab443898" containerName="collect-profiles" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.165452 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="7259e62c-d061-4d29-a3c3-c162bb8af311" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.165469 4914 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b806fa9-949d-44d4-8eb7-22a2e7f97a92" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.166218 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.178602 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29397061-p7j6w"] Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.269329 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-fernet-keys\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.269400 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8627x\" (UniqueName: \"kubernetes.io/projected/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-kube-api-access-8627x\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.269465 4914 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-config-data\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.370344 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8627x\" (UniqueName: \"kubernetes.io/projected/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-kube-api-access-8627x\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.370466 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-config-data\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.370574 4914 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-fernet-keys\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.378214 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-config-data\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.378557 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-fernet-keys\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.399606 4914 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8627x\" (UniqueName: \"kubernetes.io/projected/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-kube-api-access-8627x\") pod \"keystone-cron-29397061-p7j6w\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.533542 4914 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:00 crc kubenswrapper[4914]: I1122 15:01:00.952011 4914 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29397061-p7j6w"] Nov 22 15:01:01 crc kubenswrapper[4914]: I1122 15:01:01.316345 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" event={"ID":"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0","Type":"ContainerStarted","Data":"92abd7eb7657875408e09d165a931171a0026b5732e4618736e47a20aab0b12d"} Nov 22 15:01:01 crc kubenswrapper[4914]: I1122 15:01:01.316441 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" event={"ID":"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0","Type":"ContainerStarted","Data":"bb6c16e88254c43c0f55717f844e31bf068d739b91d76e931f67041a36c9aad9"} Nov 22 15:01:01 crc kubenswrapper[4914]: I1122 15:01:01.341352 4914 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" podStartSLOduration=1.3413330270000001 podStartE2EDuration="1.341333027s" podCreationTimestamp="2025-11-22 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-22 15:01:01.337102403 +0000 UTC m=+2027.145672074" watchObservedRunningTime="2025-11-22 15:01:01.341333027 +0000 UTC m=+2027.149902688" Nov 22 15:01:03 crc kubenswrapper[4914]: I1122 15:01:03.331558 4914 generic.go:334] "Generic (PLEG): container finished" podID="ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0" containerID="92abd7eb7657875408e09d165a931171a0026b5732e4618736e47a20aab0b12d" exitCode=0 Nov 22 15:01:03 crc kubenswrapper[4914]: I1122 15:01:03.331630 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" event={"ID":"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0","Type":"ContainerDied","Data":"92abd7eb7657875408e09d165a931171a0026b5732e4618736e47a20aab0b12d"} Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.622539 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.747612 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-fernet-keys\") pod \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.747728 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8627x\" (UniqueName: \"kubernetes.io/projected/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-kube-api-access-8627x\") pod \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.747843 4914 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-config-data\") pod \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\" (UID: \"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0\") " Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.753914 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0" (UID: "ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.754377 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-kube-api-access-8627x" (OuterVolumeSpecName: "kube-api-access-8627x") pod "ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0" (UID: "ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0"). InnerVolumeSpecName "kube-api-access-8627x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.790793 4914 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-config-data" (OuterVolumeSpecName: "config-data") pod "ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0" (UID: "ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.850204 4914 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-config-data\") on node \"crc\" DevicePath \"\"" Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.850234 4914 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 22 15:01:04 crc kubenswrapper[4914]: I1122 15:01:04.850244 4914 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8627x\" (UniqueName: \"kubernetes.io/projected/ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0-kube-api-access-8627x\") on node \"crc\" DevicePath \"\"" Nov 22 15:01:05 crc kubenswrapper[4914]: I1122 15:01:05.345778 4914 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" event={"ID":"ad5b762b-1ea1-46ca-b84b-df6ff7acc0b0","Type":"ContainerDied","Data":"bb6c16e88254c43c0f55717f844e31bf068d739b91d76e931f67041a36c9aad9"} Nov 22 15:01:05 crc kubenswrapper[4914]: I1122 15:01:05.345807 4914 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29397061-p7j6w" Nov 22 15:01:05 crc kubenswrapper[4914]: I1122 15:01:05.345825 4914 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb6c16e88254c43c0f55717f844e31bf068d739b91d76e931f67041a36c9aad9" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110350074024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110350075017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110343614016503 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110343614015453 5ustar corecore